+ source /setup_export_logs.sh
++ CLICKHOUSE_CI_LOGS_CREDENTIALS=/tmp/export-logs-config.sh
++ CLICKHOUSE_CI_LOGS_USER=ci
++ CLICKHOUSE_CI_LOGS_CLUSTER=system_logs_export
++ EXTRA_COLUMNS='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, '
++ EXTRA_COLUMNS_EXPRESSION='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id'
++ EXTRA_ORDER_BY_COLUMNS=check_name
++ EXTRA_COLUMNS_TRACE_LOG='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, symbols Array(LowCardinality(String)), lines Array(LowCardinality(String)), '
++ EXTRA_COLUMNS_EXPRESSION_TRACE_LOG='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id, arrayMap(x -> demangle(addressToSymbol(x)), trace)::Array(LowCardinality(String)) AS symbols, arrayMap(x -> addressToLine(x), trace)::Array(LowCardinality(String)) AS lines'
++ EXTRA_COLUMNS_COVERAGE_LOG='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, symbols Array(LowCardinality(String)), '
++ EXTRA_COLUMNS_EXPRESSION_COVERAGE_LOG='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id, arrayDistinct(arrayMap(x -> demangle(addressToSymbol(x)), coverage))::Array(LowCardinality(String)) AS symbols'
+ source /repo/tests/docker_scripts/stress_tests.lib
++ OK='\tOK\t\N\t'
++ FAIL='\tFAIL\t\N\t'
++ FAILURE_CONTEXT_LINES=100
++ FAILURE_CONTEXT_MAX_LINE_WIDTH=300
+ dmesg --clear
+ set -e -x -a
+ USE_DATABASE_REPLICATED=0
+ USE_SHARED_CATALOG=0
++ rg -v '#' /usr/share/zoneinfo/zone.tab
++ awk '{print $3}'
++ shuf
++ head -n1
+ TZ=America/St_Lucia
Chosen random timezone America/St_Lucia
+ echo 'Chosen random timezone America/St_Lucia'
+ ln -snf /usr/share/zoneinfo/America/St_Lucia /etc/localtime
+ echo America/St_Lucia
+ dpkg -i package_folder/clickhouse-common-static_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-common-static.
(Reading database ... 48425 files and directories currently installed.)
Preparing to unpack .../clickhouse-common-static_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-common-static (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-common-static (24.12.2.20221.altinityantalya+tsan) ...
+ dpkg -i package_folder/clickhouse-common-static-dbg_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-common-static-dbg.
(Reading database ... 48452 files and directories currently installed.)
Preparing to unpack .../clickhouse-common-static-dbg_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-common-static-dbg (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-common-static-dbg (24.12.2.20221.altinityantalya+tsan) ...
+ dpkg -i package_folder/clickhouse-odbc-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-odbc-bridge.
(Reading database ... 48459 files and directories currently installed.)
Preparing to unpack .../clickhouse-odbc-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-odbc-bridge (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-odbc-bridge (24.12.2.20221.altinityantalya+tsan) ...
+ dpkg -i package_folder/clickhouse-library-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-library-bridge.
(Reading database ... 48465 files and directories currently installed.)
Preparing to unpack .../clickhouse-library-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-library-bridge (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-library-bridge (24.12.2.20221.altinityantalya+tsan) ...
+ dpkg -i package_folder/clickhouse-server_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-server.
(Reading database ... 48471 files and directories currently installed.)
Preparing to unpack .../clickhouse-server_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-server (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-server (24.12.2.20221.altinityantalya+tsan) ...
ClickHouse binary is already located at /usr/bin/clickhouse
Symlink /usr/bin/clickhouse-server already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-server to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-client to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-local to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-benchmark to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-obfuscator to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-git-import to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-compressor to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-format to /usr/bin/clickhouse.
Symlink /usr/bin/clickhouse-extract-from-config already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-extract-from-config to /usr/bin/clickhouse.
Symlink /usr/bin/clickhouse-keeper already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-keeper to /usr/bin/clickhouse.
Symlink /usr/bin/clickhouse-keeper-converter already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-keeper-converter to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-disks to /usr/bin/clickhouse.
Creating symlink /usr/bin/ch to /usr/bin/clickhouse.
Creating symlink /usr/bin/chl to /usr/bin/clickhouse.
Creating symlink /usr/bin/chc to /usr/bin/clickhouse.
Creating clickhouse group if it does not exist.
groupadd -r clickhouse
Creating clickhouse user if it does not exist.
useradd -r --shell /bin/false --home-dir /nonexistent -g clickhouse clickhouse
Will set ulimits for clickhouse user in /etc/security/limits.d/clickhouse.conf.
Creating config directory /etc/clickhouse-server/config.d that is used for tweaks of main server configuration.
Creating config directory /etc/clickhouse-server/users.d that is used for tweaks of users configuration.
Config file /etc/clickhouse-server/config.xml already exists, will keep it and extract path info from it.
/etc/clickhouse-server/config.xml has /var/lib/clickhouse/ as data path.
/etc/clickhouse-server/config.xml has /var/log/clickhouse-server/ as log path.
Users config file /etc/clickhouse-server/users.xml already exists, will keep it and extract users info from it.
Log directory /var/log/clickhouse-server/ already exists.
Creating data directory /var/lib/clickhouse/.
Creating pid directory /var/run/clickhouse-server.
chown -R clickhouse:clickhouse '/var/log/clickhouse-server/'
chown -R clickhouse:clickhouse '/var/run/clickhouse-server'
chown clickhouse:clickhouse '/var/lib/clickhouse/'
groupadd -r clickhouse-bridge
useradd -r --shell /bin/false --home-dir /nonexistent -g clickhouse-bridge clickhouse-bridge
chown -R clickhouse-bridge:clickhouse-bridge '/usr/bin/clickhouse-odbc-bridge'
chown -R clickhouse-bridge:clickhouse-bridge '/usr/bin/clickhouse-library-bridge'
Password for the default user is an empty string. See /etc/clickhouse-server/users.xml and /etc/clickhouse-server/users.d to change it.
Setting capabilities for clickhouse binary. This is optional.
chown -R clickhouse:clickhouse '/etc/clickhouse-server'
ClickHouse has been successfully installed.
Start clickhouse-server with:
sudo clickhouse start
Start clickhouse-client with:
clickhouse-client
+ dpkg -i package_folder/clickhouse-client_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-client.
(Reading database ... 48488 files and directories currently installed.)
Preparing to unpack .../clickhouse-client_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-client (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-client (24.12.2.20221.altinityantalya+tsan) ...
+ echo ''
+ [[ -z '' ]]
+ ch --query 'SELECT 1'
1
+ chl --query 'SELECT 1'
1
+ chc --version
ClickHouse client version 24.12.2.20221.altinityantalya (altinity build).
+ ln -sf /repo/tests/clickhouse-test /usr/bin/clickhouse-test
+ export CLICKHOUSE_GRPC_CLIENT=/repo/utils/grpc-client/clickhouse-grpc-client.py
+ CLICKHOUSE_GRPC_CLIENT=/repo/utils/grpc-client/clickhouse-grpc-client.py
+ source /repo/tests/docker_scripts/attach_gdb.lib
++ source /repo/tests/docker_scripts/utils.lib
+++ sysctl kernel.core_pattern=core.%e.%p-%P
kernel.core_pattern = core.%e.%p-%P
+++ sysctl fs.suid_dumpable=1
fs.suid_dumpable = 1
+ source /repo/tests/docker_scripts/utils.lib
++ sysctl kernel.core_pattern=core.%e.%p-%P
kernel.core_pattern = core.%e.%p-%P
++ sysctl fs.suid_dumpable=1
fs.suid_dumpable = 1
+ /repo/tests/config/install.sh
+ DEST_SERVER_PATH=/etc/clickhouse-server
+ DEST_CLIENT_PATH=/etc/clickhouse-client
+++ dirname /repo/tests/config/install.sh
++ cd /repo/tests/config
++ pwd -P
+ SRC_PATH=/repo/tests/config
+ '[' 0 -ge 2 ']'
+ FAST_TEST=0
+ NO_AZURE=0
+ [[ 0 -gt 0 ]]
+ echo 'Going to install test configs from /repo/tests/config into /etc/clickhouse-server'
+ mkdir -p /etc/clickhouse-server/config.d/
Going to install test configs from /repo/tests/config into /etc/clickhouse-server
+ mkdir -p /etc/clickhouse-server/users.d/
+ mkdir -p /etc/clickhouse-client
+ ln -sf /repo/tests/config/config.d/zookeeper_write.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/max_num_to_warn.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/listen.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/text_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/blob_storage_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/custom_settings_prefixes.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/database_catalog_drop_table_concurrency.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/enable_access_control_improvements.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/macros.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/secure_ports.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/clusters.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/graphite.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/graphite_alternative.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/grpc_protocol.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/database_atomic.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/max_concurrent_queries.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/merge_tree_settings.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/backoff_failed_mutation.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/merge_tree_old_dirs_cleanup.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/test_cluster_with_incorrect_pw.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/keeper_port.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/logging_no_rotate.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/merge_tree.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/lost_forever_check.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/tcp_with_proxy.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/prometheus.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/top_level_domains_lists.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/top_level_domains_path.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/transactions.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/encryption.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/CORS.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/zookeeper_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/logger_trace.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/named_collection.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/ssl_certs.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/filesystem_cache_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/session_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/system_unfreeze.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/enable_zero_copy_replication.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/nlp.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/forbidden_headers.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/enable_keeper_map.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/custom_disks_base_path.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/display_name.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/compressed_marks_and_index.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/disable_s3_env_credentials.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/enable_wait_for_shutdown_replicated_tables.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/backups.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/filesystem_caches_path.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/validate_tcp_client_information.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/zero_copy_destructive_operations.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/block_number.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/handlers.yaml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/serverwide_trace_collector.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/rocksdb.xml /etc/clickhouse-server/config.d/
+ '[' 0 '!=' 1 ']'
+ ln -sf /repo/tests/config/config.d/legacy_geobase.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/users.d/log_queries.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/readonly.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/access_management.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/database_atomic_drop_detach_sync.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/opentelemetry.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/remote_queries.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/session_log_test.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/memory_profiler.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/no_fsync_metadata.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/filelog.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/enable_blobs_check.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/marks.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/insert_keeper_retries.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/prefetch_settings.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/nonconst_timezone.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/allow_introspection_functions.yaml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/replicated_ddl_entry.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/limits.yaml /etc/clickhouse-server/users.d/
+ [[ -n '' ]]
+ ln -sf /repo/tests/config/users.d/timeouts.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/ints_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/strings_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/decimals_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/executable_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/executable_pool_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/test_function.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/top_level_domains /etc/clickhouse-server/
+ ln -sf /repo/tests/config/regions_hierarchy.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/regions_names_en.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/regions_names_es.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/ext-en.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/ext-ru.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/lem-en.bin /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/server.key /etc/clickhouse-server/
+ ln -sf /repo/tests/config/server.crt /etc/clickhouse-server/
+ ln -sf /repo/tests/config/dhparam.pem /etc/clickhouse-server/
+ ln -sf --backup=simple --suffix=_original.xml /repo/tests/config/config.d/query_masking_rules.xml /etc/clickhouse-server/config.d/
+ [[ -n '' ]]
+ rm -f /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml
+ ln -sf /repo/tests/config/config.d/zookeeper.xml /etc/clickhouse-server/config.d/
+ [[ -n '' ]]
+ rm -f /etc/clickhouse-server/config.d/cannot_allocate_thread_injection.xml
+ value=1
+ sed --follow-symlinks -i 's|[01]|1|' /etc/clickhouse-server/config.d/keeper_port.xml
+ value=67047424
+ sed --follow-symlinks -i 's|[[:digit:]]\+|67047424|' /etc/clickhouse-server/config.d/keeper_port.xml
+ value=43378688
+ sed --follow-symlinks -i 's|[[:digit:]]\+|43378688|' /etc/clickhouse-server/config.d/keeper_port.xml
+ [[ -n '' ]]
+ [[ -n '' ]]
+ [[ '' == \1 ]]
+ [[ '' == \1 ]]
+ [[ 1 == \1 ]]
+ [[ 0 != \1 ]]
+ ln -sf /repo/tests/config/config.d/azure_storage_conf.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/storage_conf.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/storage_conf_02944.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/storage_conf_02963.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/storage_conf_02961.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/users.d/s3_cache.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/s3_cache_new.xml /etc/clickhouse-server/users.d/
+ [[ 0 == \1 ]]
+ ln -sf /repo/tests/config/client_config.xml /etc/clickhouse-client/config.xml
+ /repo/tests/docker_scripts/setup_minio.sh stateless
+ azurite-blob --blobHost 0.0.0.0 --blobPort 10000 --silent --inMemoryPersistence
+ export MINIO_ROOT_USER=clickhouse
+ MINIO_ROOT_USER=clickhouse
+ export MINIO_ROOT_PASSWORD=clickhouse
+ MINIO_ROOT_PASSWORD=clickhouse
+ main stateless
+ local query_dir
++ check_arg stateless
++ local query_dir
++ '[' '!' 1 -eq 1 ']'
++ case "$1" in
++ query_dir=0_stateless
++ echo 0_stateless
+ query_dir=0_stateless
+ '[' '!' -f ./minio ']'
+ start_minio
+ mkdir -p ./minio_data
+ ./minio --version
minio version RELEASE.2024-08-03T04-33-23Z (commit-id=6efb56851c40da88d1ca15112e2d686a4ecec6b3)
Runtime: go1.22.5 linux/amd64
License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html
Copyright: 2015-2024 MinIO, Inc.
+ wait_for_it
+ local counter=0
+ local max_counter=60
+ local url=http://localhost:11111
+ ./minio server --address :11111 ./minio_data
+ params=('--silent' '--verbose')
+ local params
+ curl --silent --verbose http://localhost:11111
+ grep AccessDenied
trying to connect to minio
+ [[ 0 == \6\0 ]]
+ echo 'trying to connect to minio'
+ sleep 1
(node:273) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead.
(Use `node --trace-deprecation ...` to show where the warning was created)
In-memory extent storage is enabled with a limit of 15667.21 MB (16428261376 bytes, 50% of total memory).
Azurite Blob service is starting on 0.0.0.0:10000
Azurite Blob service successfully listens on http://0.0.0.0:10000
INFO: Formatting 1st pool, 1 set(s), 1 drives per set.
INFO: WARNING: Host local has more than 0 drives of set. A host failure will result in data becoming unavailable.
MinIO Object Storage Server
Copyright: 2015-2025 MinIO, Inc.
License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html
Version: RELEASE.2024-08-03T04-33-23Z (go1.22.5 linux/amd64)
API: http://172.17.0.2:11111 http://127.0.0.1:11111
WebUI: http://172.17.0.2:42161 http://127.0.0.1:42161
Docs: https://min.io/docs/minio/linux/index.html
+ counter=1
+ curl --silent --verbose http://localhost:11111
+ grep AccessDenied
AccessDenied
Access Denied./18324FA5B7C0C8967dc7eb22d3288ec80374614e9088e31d3668a6922ead55932dd2a8e56373820f
+ lsof -i :11111
COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME
minio 296 root 8u IPv4 32994 0t0 TCP localhost:11111 (LISTEN)
minio 296 root 9u IPv6 32995 0t0 TCP *:11111 (LISTEN)
minio 296 root 10u IPv6 32996 0t0 TCP localhost:11111 (LISTEN)
+ sleep 5
+ setup_minio stateless
+ local test_type=stateless
+ ./mc alias set clickminio http://localhost:11111 clickhouse clickhouse
Added `clickminio` successfully.
+ ./mc admin user add clickminio test testtest
Added user `test` successfully.
+ ./mc admin policy attach clickminio readwrite --user=test
Attached Policies: [readwrite]
To User: test
+ ./mc mb --ignore-existing clickminio/test
Bucket created successfully `clickminio/test`.
+ '[' stateless = stateless ']'
+ ./mc anonymous set public clickminio/test
Access permission for `clickminio/test` is set to `public`
+ upload_data 0_stateless /repo/tests/
+ local query_dir=0_stateless
+ local test_path=/repo/tests/
+ local data_path=/repo/tests//queries/0_stateless/data_minio
+ '[' -d /repo/tests//queries/0_stateless/data_minio ']'
+ ./mc cp --recursive /repo/tests//queries/0_stateless/data_minio/ clickminio/test/
`/repo/tests/queries/0_stateless/data_minio/02731.parquet` -> `clickminio/test/02731.parquet`
`/repo/tests/queries/0_stateless/data_minio/02876.parquet` -> `clickminio/test/02876.parquet`
`/repo/tests/queries/0_stateless/data_minio/02366_data.jsonl` -> `clickminio/test/02366_data.jsonl`
`/repo/tests/queries/0_stateless/data_minio/02731.arrow` -> `clickminio/test/02731.arrow`
`/repo/tests/queries/0_stateless/data_minio/03036_archive1.tar` -> `clickminio/test/03036_archive1.tar`
`/repo/tests/queries/0_stateless/data_minio/03036_archive1.zip` -> `clickminio/test/03036_archive1.zip`
`/repo/tests/queries/0_stateless/data_minio/03036_archive2.tar` -> `clickminio/test/03036_archive2.tar`
`/repo/tests/queries/0_stateless/data_minio/03036_archive2.zip` -> `clickminio/test/03036_archive2.zip`
`/repo/tests/queries/0_stateless/data_minio/03036_archive3.tar.gz` -> `clickminio/test/03036_archive3.tar.gz`
`/repo/tests/queries/0_stateless/data_minio/03036_compressed_file_archive.zip` -> `clickminio/test/03036_compressed_file_archive.zip`
`/repo/tests/queries/0_stateless/data_minio/03036_json_archive.zip` -> `clickminio/test/03036_json_archive.zip`
`/repo/tests/queries/0_stateless/data_minio/a.tsv` -> `clickminio/test/a.tsv`
`/repo/tests/queries/0_stateless/data_minio/b.tsv` -> `clickminio/test/b.tsv`
`/repo/tests/queries/0_stateless/data_minio/c.tsv` -> `clickminio/test/c.tsv`
`/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/column1=Gordon/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/column1=Gordon/sample.parquet`
`/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/column1=Schmidt/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/column1=Schmidt/sample.parquet`
`/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/sample.parquet`
`/repo/tests/queries/0_stateless/data_minio/hive_partitioning/non_existing_column=Elizabeth/sample.parquet` -> `clickminio/test/hive_partitioning/non_existing_column=Elizabeth/sample.parquet`
`/repo/tests/queries/0_stateless/data_minio/json_data` -> `clickminio/test/json_data`
`/repo/tests/queries/0_stateless/data_minio/tsv_with_header.tsv` -> `clickminio/test/tsv_with_header.tsv`
Total: 5.42 MiB, Transferred: 5.42 MiB, Speed: 128.74 MiB/s
+ setup_aws_credentials
+ local minio_root_user=clickhouse
+ local minio_root_password=clickhouse
+ mkdir -p /root/.aws
+ cat
+ config_logs_export_cluster /etc/clickhouse-server/config.d/system_logs_export.yaml
+ set +x
File /tmp/export-logs-config.sh does not exist, do not setup
+ [[ -n '' ]]
+ export IS_FLAKY_CHECK=0
+ IS_FLAKY_CHECK=0
+ export NUM_TRIES
+ '[' 1 -gt 1 ']'
+ sudo -E -u clickhouse /usr/bin/clickhouse-server --config /etc/clickhouse-server/config.xml --daemon --pid-file /var/run/clickhouse-server/clickhouse-server.pid
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for _ in {1..100}
+ clickhouse-client --query 'SELECT 1'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ sleep 1
+ for _ in {1..100}
+ clickhouse-client --query 'SELECT 1'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ sleep 1
+ for _ in {1..100}
+ clickhouse-client --query 'SELECT 1'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ sleep 1
+ for _ in {1..100}
+ clickhouse-client --query 'SELECT 1'
1
+ break
+ setup_logs_replication
+ set +x
File /tmp/export-logs-config.sh does not exist, do not setup
+ attach_gdb_to_clickhouse
++ run_with_retry 5 clickhouse-client --query 'SELECT count() FROM system.build_options WHERE name = '\''CXX_FLAGS'\'' AND position('\''sanitize=address'\'' IN value)'
++ [[ ahxB =~ e ]]
++ set_e=false
++ set +e
++ local total_retries=5
++ shift
++ local retry=0
++ '[' 0 -ge 5 ']'
++ clickhouse-client --query 'SELECT count() FROM system.build_options WHERE name = '\''CXX_FLAGS'\'' AND position('\''sanitize=address'\'' IN value)'
++ false
++ return
+ IS_ASAN=0
+ [[ 0 = \1 ]]
++ kill -l SIGRTMIN
+ RTMIN=34
+ echo '
set follow-fork-mode parent
handle SIGHUP nostop noprint pass
handle SIGINT nostop noprint pass
handle SIGQUIT nostop noprint pass
handle SIGPIPE nostop noprint pass
handle SIGTERM nostop noprint pass
handle SIGUSR1 nostop noprint pass
handle SIGUSR2 nostop noprint pass
handle SIG34 nostop noprint pass
info signals
continue
backtrace full
info registers
p top' 1 KiB of the 'stack:
p/x *(uint64_t[128]*)$sp
maintenance info sections
thread apply all backtrace full
disassemble /s
up
disassemble /s
up
disassemble /s
p "done"
detach
quit
'
+ sleep 5
+ ts '%Y-%m-%d %H:%M:%S'
++ cat /var/run/clickhouse-server/clickhouse-server.pid
+ gdb -batch -command script.gdb -p 428
aarch64-binfmt-P: Could not open '/lib/ld-linux-aarch64.so.1': No such file or directory
+ run_with_retry 60 clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\'''
+ [[ aehxB =~ e ]]
+ set_e=true
+ set +e
+ local total_retries=60
+ shift
+ local retry=0
+ '[' 0 -ge 60 ']'
+ clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\'''
Connected to clickhouse-server after attaching gdb
+ true
+ set -e
+ return
+ clickhouse-client --allow_experimental_json_type=1 --query 'CREATE TABLE minio_audit_logs
(
log JSON(time DateTime64(9))
)
ENGINE = MergeTree
ORDER BY tuple()'
+ clickhouse-client --allow_experimental_json_type=1 --query 'CREATE TABLE minio_server_logs
(
log JSON(time DateTime64(9))
)
ENGINE = MergeTree
ORDER BY tuple()'
+ ./mc admin config set clickminio logger_webhook:ch_server_webhook 'endpoint=http://localhost:8123/?async_insert=1&wait_for_async_insert=0&async_insert_busy_timeout_min_ms=5000&async_insert_busy_timeout_max_ms=5000&async_insert_max_query_number=1000&async_insert_max_data_size=10485760&date_time_input_format=best_effort&query=INSERT%20INTO%20minio_server_logs%20FORMAT%20JSONAsObject' queue_size=1000000 batch_size=500
Successfully applied new settings.
+ ./mc admin config set clickminio audit_webhook:ch_audit_webhook 'endpoint=http://localhost:8123/?async_insert=1&wait_for_async_insert=0&async_insert_busy_timeout_min_ms=5000&async_insert_busy_timeout_max_ms=5000&async_insert_max_query_number=1000&async_insert_max_data_size=10485760&date_time_input_format=best_effort&query=INSERT%20INTO%20minio_audit_logs%20FORMAT%20JSONAsObject' queue_size=1000000 batch_size=500
Successfully applied new settings.
+ max_retries=100
+ retry=1
+ '[' 1 -le 100 ']'
+ echo 'clickminio restart attempt 1:'
clickminio restart attempt 1:
++ ./mc admin service restart clickminio --wait --json
++ jq -r .status
INFO: Restarting on service signal
MinIO Object Storage Server
Copyright: 2015-2025 MinIO, Inc.
License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html
Version: RELEASE.2024-08-03T04-33-23Z (go1.22.5 linux/amd64)
API: http://172.17.0.2:11111 http://127.0.0.1:11111
WebUI: http://172.17.0.2:36665 http://127.0.0.1:36665
Docs: https://min.io/docs/minio/linux/index.html
Output of restart status: success
success
Restarted clickminio successfully.
+ output='success
success'
+ echo 'Output of restart status: success
success'
+ expected_output='success
success'
+ '[' 'success
success' = 'success
success' ']'
+ echo 'Restarted clickminio successfully.'
+ break
+ '[' 1 -gt 100 ']'
+ MC_ADMIN_PID=1344
+ ./mc admin trace clickminio
+ export -f run_tests
+ '[' 1 -gt 1 ']'
+ run_tests
+ set -x
+ read -ra ADDITIONAL_OPTIONS
+ HIGH_LEVEL_COVERAGE=YES
+ '[' 1 -gt 1 ']'
+ [[ -n '' ]]
+ [[ -n '' ]]
+ [[ 0 -eq 1 ]]
+ [[ '' -eq 1 ]]
+ [[ 0 -eq 1 ]]
++ clickhouse-client --query 'SELECT value LIKE '\''%SANITIZE_COVERAGE%'\'' FROM system.build_options WHERE name = '\''CXX_FLAGS'\'''
+ [[ 1 == 0 ]]
+ ADDITIONAL_OPTIONS+=('--jobs')
+ ADDITIONAL_OPTIONS+=('8')
+ [[ -n 2 ]]
+ [[ -n 4 ]]
+ ADDITIONAL_OPTIONS+=('--run-by-hash-num')
+ ADDITIONAL_OPTIONS+=("$RUN_BY_HASH_NUM")
+ ADDITIONAL_OPTIONS+=('--run-by-hash-total')
+ ADDITIONAL_OPTIONS+=("$RUN_BY_HASH_TOTAL")
+ HIGH_LEVEL_COVERAGE=NO
+ [[ -n '' ]]
+ [[ NO = \Y\E\S ]]
+ ADDITIONAL_OPTIONS+=('--report-logs-stats')
+ try_run_with_retry 10 clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')'
+ local total_retries=10
+ shift
+ fn_exists run_with_retry
+ declare -F run_with_retry
+ run_with_retry 10 clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')'
+ [[ aehxB =~ e ]]
+ set_e=true
+ set +e
+ local total_retries=10
+ shift
+ local retry=0
+ '[' 0 -ge 10 ']'
+ clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')'
+ true
+ set -e
+ return
+ set +e
+ TEST_ARGS=(--testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries "/repo/tests/queries" --test-runs "$NUM_TRIES" "${ADDITIONAL_OPTIONS[@]}")
+ clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 2 --run-by-hash-total 4 --report-logs-stats
+ ts '%Y-%m-%d %H:%M:%S'
+ tee -a test_output/test_result.txt
2025-04-01 17:39:56 Using queries from '/repo/tests/queries' directory
2025-04-01 17:39:56 Connecting to ClickHouse server... OK
2025-04-01 17:39:56 Connected to server 24.12.2.20221.altinityantalya @ 82252d159dc02cab0f366aaa5691adc1545dd11d HEAD
2025-04-01 17:39:56 Found 1667 parallel tests and 130 sequential tests
2025-04-01 17:39:57 Running about 208 stateless tests (Process-8).
2025-04-01 17:39:57 03085_analyzer_alias_column_group_by: [ OK ] 0.58 sec.
2025-04-01 17:39:57 Running about 208 stateless tests (Process-10).
2025-04-01 17:39:57 01906_partition_by_multiply_by_zero: [ OK ] 0.73 sec.
2025-04-01 17:39:57 Running about 208 stateless tests (Process-6).
2025-04-01 17:39:57 03171_direct_dict_short_circuit_bug: [ OK ] 0.83 sec.
2025-04-01 17:39:57 01851_hedged_connections_external_tables: [ SKIPPED ] 0.00 sec.
2025-04-01 17:39:57 Reason: not running for current build
2025-04-01 17:39:57 Running about 208 stateless tests (Process-5).
2025-04-01 17:39:57 01010_pm_join_all_join_bug: [ OK ] 0.89 sec.
2025-04-01 17:39:57 Running about 208 stateless tests (Process-9).
2025-04-01 17:39:57 02724_persist_interval_type: [ OK ] 0.88 sec.
2025-04-01 17:39:57 Running about 208 stateless tests (Process-3).
2025-04-01 17:39:57 00228_shard_quantiles_deterministic_merge_overflow: [ OK ] 1.44 sec.
2025-04-01 17:39:58 02815_first_line: [ OK ] 0.68 sec.
2025-04-01 17:39:58 01549_low_cardinality_mv_fuzz: [ OK ] 0.68 sec.
2025-04-01 17:39:58 00251_has_types: [ OK ] 0.78 sec.
2025-04-01 17:39:58 02372_now_in_block: [ OK ] 1.18 sec.
2025-04-01 17:39:58 00820_multiple_joins_subquery_requires_alias: [ OK ] 1.13 sec.
2025-04-01 17:39:58 01914_ubsan_quantile_timing: [ OK ] 0.53 sec.
2025-04-01 17:39:58 02815_alias_to_length: [ OK ] 0.53 sec.
2025-04-01 17:39:58 00492_drop_temporary_table: [ OK ] 0.62 sec.
2025-04-01 17:39:59 00361_shared_array_offsets_and_squash_blocks: [ OK ] 0.73 sec.
2025-04-01 17:39:59 03230_subcolumns_mv: [ OK ] 0.73 sec.
2025-04-01 17:39:59 02465_limit_trivial_max_rows_to_read: [ OK ] 0.88 sec.
2025-04-01 17:40:00 Running about 208 stateless tests (Process-4).
2025-04-01 17:40:00 03230_keeper_cp_mv_commands: [ OK ] 3.49 sec.
2025-04-01 17:40:00 02723_param_exception_message_context: [ OK ] 2.13 sec.
2025-04-01 17:40:01 01043_h3_edge_length_m: [ OK ] 0.58 sec.
2025-04-01 17:40:01 00545_weird_aggregate_functions: [ OK ] 0.58 sec.
2025-04-01 17:40:02 01196_max_parser_depth: [ OK ] 2.63 sec.
2025-04-01 17:40:02 02352_interactive_queries_from_file: [ OK ] 2.13 sec.
2025-04-01 17:40:02 01436_storage_merge_with_join_push_down: [ OK ] 0.83 sec.
2025-04-01 17:40:02 Running about 208 stateless tests (Process-7).
2025-04-01 17:40:02 03031_minmax_index_for_pointinpolygon: [ OK ] 6.29 sec.
2025-04-01 17:40:02 00276_sample: [ OK ] 4.48 sec.
2025-04-01 17:40:03 02012_sha512_fixedstring: [ OK ] 0.63 sec.
2025-04-01 17:40:03 01036_union_different_columns: [ OK ] 0.68 sec.
2025-04-01 17:40:04 03285_analyzer_extract_common_expr_bug: [ OK ] 0.63 sec.
2025-04-01 17:40:04 02250_lots_of_columns_in_csv_with_names: [ OK ] 4.99 sec.
2025-04-01 17:40:05 03237_get_subcolumn_low_cardinality_column: [ OK ] 0.53 sec.
2025-04-01 17:40:05 02157_readonly_system_suspend: [ OK ] 1.98 sec.
2025-04-01 17:40:06 00122_join_with_subquery_with_subquery: [ OK ] 0.58 sec.
2025-04-01 17:40:06 01505_trivial_count_with_partition_predicate: [ OK ] 1.18 sec.
2025-04-01 17:40:06 03262_analyzer_materialized_view_in_with_cte: [ OK ] 0.73 sec.
2025-04-01 17:40:06 03198_unload_primary_key_outdated: [ OK ] 4.69 sec.
2025-04-01 17:40:07 00396_uuid: [ OK ] 0.73 sec.
2025-04-01 17:40:08 01949_heredoc_unfinished: [ OK ] 1.64 sec.
2025-04-01 17:40:08 01601_proxy_protocol: [ OK ] 1.48 sec.
2025-04-01 17:40:09 00365_statistics_in_formats: [ OK ] 5.04 sec.
2025-04-01 17:40:09 01214_test_storage_merge_aliases_with_where: [ OK ] 1.13 sec.
2025-04-01 17:40:10 02366_window_function_order_by: [ OK ] 0.68 sec.
2025-04-01 17:40:10 00271_agg_state_and_totals: [ OK ] 0.68 sec.
2025-04-01 17:40:10 03096_largest_triangle_3b_crash: [ OK ] 0.53 sec.
2025-04-01 17:40:11 03064_analyzer_named_subqueries: [ OK ] 0.58 sec.
2025-04-01 17:40:11 02158_proportions_ztest_cmp: [ OK ] 2.68 sec.
2025-04-01 17:40:11 01866_datetime64_cmp_with_constant: [ OK ] 1.08 sec.
2025-04-01 17:40:11 02843_backup_use_same_password_for_base_backup: [ OK ] 9.55 sec.
2025-04-01 17:40:11 03263_analyzer_materialized_view_cte_nested: [ OK ] 0.73 sec.
2025-04-01 17:40:13 02122_parallel_formatting_JSONCompactEachRowWithNames: [ OK ] 6.75 sec.
2025-04-01 17:40:14 02293_grouping_function_group_by: [ OK ] 2.54 sec.
2025-04-01 17:40:14 01851_clear_column_referenced_by_mv: [ OK ] 0.78 sec.
2025-04-01 17:40:14 00652_replicated_mutations_default_database_zookeeper: [ OK ] 3.69 sec.
2025-04-01 17:40:15 01416_join_totals_header_bug: [ OK ] 0.73 sec.
2025-04-01 17:40:15 01825_type_json_18: [ OK ] 0.64 sec.
2025-04-01 17:40:16 02538_alter_rename_sequence: [ OK ] 1.43 sec.
2025-04-01 17:40:18 01018_Distributed__shard_num: [ OK ] 2.39 sec.
2025-04-01 17:40:18 01289_min_execution_speed_not_too_early: [ OK ] 15.47 sec.
2025-04-01 17:40:18 00823_sequence_match_dfa: [ OK ] 6.40 sec.
2025-04-01 17:40:19 02515_and_or_if_multiif_not_return_lc: [ OK ] 0.57 sec.
2025-04-01 17:40:19 01769_extended_range_2: [ OK ] 0.83 sec.
2025-04-01 17:40:19 00050_any_left_join: [ OK ] 0.58 sec.
2025-04-01 17:40:19 03165_storage_merge_view_prewhere: [ OK ] 0.78 sec.
2025-04-01 17:40:20 03068_analyzer_distributed_join: [ OK ] 0.88 sec.
2025-04-01 17:40:21 02433_default_expression_operator_in: [ OK ] 1.08 sec.
2025-04-01 17:40:22 02861_alter_replace_partition_do_not_wait_mutations_on_unrelated_partitions: [ OK ] 7.09 sec.
2025-04-01 17:40:22 02932_group_by_null_fuzzer: [ OK ] 0.63 sec.
2025-04-01 17:40:22 00369_int_div_of_float: [ OK ] 0.58 sec.
2025-04-01 17:40:22 01455_nullable_type_with_if_agg_combinator: [ OK ] 0.58 sec.
2025-04-01 17:40:23 02916_replication_protocol_wait_for_part: [ OK ] 11.20 sec.
2025-04-01 17:40:23 01762_deltasumtimestamp: [ OK ] 0.73 sec.
2025-04-01 17:40:24 03033_dist_settings.optimize_skip_unused_shards_rewrite_in_composite_sharding_key: [ OK ] 0.78 sec.
2025-04-01 17:40:24 01273_arrow_nested_arrays_load: [ OK ] 4.58 sec.
2025-04-01 17:40:25 01303_polygons_equals: [ OK ] 0.63 sec.
2025-04-01 17:40:26 00969_columns_clause: [ OK ] 1.38 sec.
2025-04-01 17:40:26 00752_low_cardinality_permute: [ OK ] 0.89 sec.
2025-04-01 17:40:26 02516_projections_with_rollup: [ OK ] 28.20 sec.
2025-04-01 17:40:26 01921_with_fill_with_totals: [ OK ] 0.68 sec.
2025-04-01 17:40:27 02158_ztest: [ OK ] 0.70 sec.
2025-04-01 17:40:27 00900_orc_arrow_parquet_tuples: [ OK ] 9.30 sec.
2025-04-01 17:40:27 02723_jit_aggregation_bug_48120: [ OK ] 0.93 sec.
2025-04-01 17:40:28 00619_union_highlite: [ OK ] 0.79 sec.
2025-04-01 17:40:29 02534_parquet_fixed_binary_array: [ OK ] 6.19 sec.
2025-04-01 17:40:30 01045_array_zip: [ OK ] 1.24 sec.
2025-04-01 17:40:30 01477_lc_in_merge_join_left_key: [ OK ] 2.74 sec.
2025-04-01 17:40:31 01413_alter_update_supertype: [ OK ] 1.29 sec.
2025-04-01 17:40:32 02813_seriesDecomposeSTL: [ OK ] 1.75 sec.
2025-04-01 17:40:32 02024_compression_in_query: [ OK ] 6.82 sec.
2025-04-01 17:40:32 02985_if_over_big_int_decimal: [ OK ] 0.94 sec.
2025-04-01 17:40:33 00942_mv_rename_table: [ OK ] 0.84 sec.
2025-04-01 17:40:33 02113_untuple_func_alias: [ OK ] 0.59 sec.
2025-04-01 17:40:34 02030_function_mapContainsKeyLike: [ OK ] 1.12 sec.
2025-04-01 17:40:34 00007_array: [ OK ] 0.59 sec.
2025-04-01 17:40:35 02212_h3_get_pentagon_indexes: [ OK ] 1.09 sec.
2025-04-01 17:40:35 02129_add_column_add_ttl: [ OK ] 2.17 sec.
2025-04-01 17:40:36 03034_recursive_cte_tree: [ OK ] 0.89 sec.
2025-04-01 17:40:36 00723_remerge_sort: [ OK ] 8.70 sec.
2025-04-01 17:40:37 02733_sparse_columns_reload: [ OK ] 1.04 sec.
2025-04-01 17:40:37 00380_client_break_at_exception_in_batch_mode: [ OK ] 2.14 sec.
2025-04-01 17:40:38 02496_remove_redundant_sorting_analyzer: [ OK ] 22.09 sec.
2025-04-01 17:40:40 02952_clickhouse_local_query_parameters_cli: [ OK ] 1.98 sec.
2025-04-01 17:40:41 00700_to_decimal_or_something_1: [ OK ] 4.50 sec.
2025-04-01 17:40:42 02751_text_formats_bad_nullable_parsing: [ OK ] 5.55 sec.
2025-04-01 17:40:43 02135_local_create_db: [ OK ] 2.24 sec.
2025-04-01 17:40:44 02122_4letter_words_stress_zookeeper: [ OK ] 21.44 sec.
2025-04-01 17:40:44 01551_mergetree_read_in_order_spread: [ OK ] 0.88 sec.
2025-04-01 17:40:45 02870_per_column_settings: [ OK ] 1.23 sec.
2025-04-01 17:40:46 02316_hierarchical_dictionaries_nullable_parent_key: [ OK ] 1.93 sec.
2025-04-01 17:40:47 01062_pm_all_join_with_block_continuation: [ OK ] 48.31 sec.
2025-04-01 17:40:49 02498_random_string_in_json_schema_inference: [ OK ] 1.93 sec.
2025-04-01 17:40:52 02149_schema_inference_create_table_syntax: [ OK ] 9.80 sec.
2025-04-01 17:40:53 01499_log_deadlock: [ OK ] 0.74 sec.
2025-04-01 17:40:54 01655_plan_optimizations: [ OK ] 26.02 sec.
2025-04-01 17:40:54 00666_uniq_complex_types: [ OK ] 1.18 sec.
2025-04-01 17:40:55 02999_scalar_subqueries_bug_2: [ OK ] 0.58 sec.
2025-04-01 17:40:56 02540_date_column_consistent_insert_behaviour: [ OK ] 1.88 sec.
2025-04-01 17:40:56 02864_statistics_predicates: [ OK ] 11.27 sec.
2025-04-01 17:40:57 02295_GROUP_BY_AggregateFunction: [ OK ] 0.94 sec.
2025-04-01 17:40:57 00157_aliases_and_lambda_formal_parameters: [ OK ] 0.48 sec.
2025-04-01 17:40:58 01938_joins_identifiers: [ OK ] 0.73 sec.
2025-04-01 17:40:58 01505_pipeline_executor_UAF: [ OK ] 23.82 sec.
2025-04-01 17:40:59 02311_range_hashed_dictionary_range_cast: [ OK ] 0.69 sec.
2025-04-01 17:40:59 03036_reading_s3_archives: [ OK ] 2.39 sec.
2025-04-01 17:40:59 03046_column_in_block_array_join: [ OK ] 0.68 sec.
2025-04-01 17:40:59 00647_histogram: [ OK ] 0.68 sec.
2025-04-01 17:40:59 00581_limit_on_result_and_subquery_and_insert: [ OK ] 0.63 sec.
2025-04-01 17:40:59 02766_bitshift_with_const_arguments: [ OK ] 0.83 sec.
2025-04-01 17:41:00 03205_system_sync_replica_format: [ OK ] 0.58 sec.
2025-04-01 17:41:00 03071_analyzer_array_join_forbid_non_existing_columns: [ OK ] 0.79 sec.
2025-04-01 17:41:01 03162_dynamic_type_nested: [ OK ] 0.63 sec.
2025-04-01 17:41:01 01458_named_tuple_millin: [ OK ] 0.53 sec.
2025-04-01 17:41:01 03257_async_insert_native_empty_block: [ OK ] 6.69 sec.
2025-04-01 17:41:01 03224_json_merges_new_type_in_shared_data: [ OK ] 0.78 sec.
2025-04-01 17:41:01 03270_processors_profile_log_3: [ OK ] 12.41 sec.
2025-04-01 17:41:02 01010_partial_merge_join_const_and_lc: [ OK ] 0.78 sec.
2025-04-01 17:41:04 01780_column_sparse_full: [ OK ] 2.68 sec.
2025-04-01 17:41:04 01339_client_unrecognized_option: [ OK ] 2.34 sec.
2025-04-01 17:41:04 02020_cast_integer_overflow: [ OK ] 0.53 sec.
2025-04-01 17:41:04 00003_reinterpret_as_string: [ OK ] 0.53 sec.
2025-04-01 17:41:05 02294_dictionaries_hierarchical_index: [ OK ] 1.08 sec.
2025-04-01 17:41:06 01214_point_in_Mecca: [ OK ] 4.24 sec.
2025-04-01 17:41:06 02499_analyzer_set_index: [ OK ] 0.68 sec.
2025-04-01 17:41:07 01926_date_date_time_supertype: [ OK ] 0.68 sec.
2025-04-01 17:41:08 01825_type_json_missed_values: [ OK ] 1.18 sec.
2025-04-01 17:41:10 00339_parsing_bad_arrays: [ OK ] 1.48 sec.
2025-04-01 17:41:10 01098_msgpack_format: [ OK ] 29.96 sec.
2025-04-01 17:41:10 02896_leading_zeroes_no_octal: [ OK ] 4.33 sec.
2025-04-01 17:41:11 02513_prewhere_combine_step_filters: [ OK ] 1.13 sec.
2025-04-01 17:41:11 02508_index_analysis_to_date_timezone: [ OK ] 0.83 sec.
2025-04-01 17:41:11 00030_alter_table: [ OK ] 1.48 sec.
2025-04-01 17:41:12 02788_current_schemas_function: [ OK ] 0.88 sec.
2025-04-01 17:41:12 03024_total_rows_approx_is_set_for_system_zeros_and_generate_random: [ OK ] 0.73 sec.
2025-04-01 17:41:12 02864_statistics_delayed_materialization_in_merge: [ OK ] 1.08 sec.
2025-04-01 17:41:13 03243_array_join_lambda: [ OK ] 0.58 sec.
2025-04-01 17:41:13 03203_variant_convert_field_to_type_bug: [ OK ] 0.63 sec.
2025-04-01 17:41:13 03239_if_constant_folding: [ OK ] 0.74 sec.
2025-04-01 17:41:13 02534_join_prewhere_bug: [ OK ] 0.88 sec.
2025-04-01 17:41:14 01846_alter_column_without_type_bugfix: [ OK ] 0.68 sec.
2025-04-01 17:41:14 00552_logical_functions_uint8_as_bool: [ OK ] 0.68 sec.
2025-04-01 17:41:15 02324_map_combinator_bug: [ OK ] 1.13 sec.
2025-04-01 17:41:16 01710_projection_with_column_transformers: [ OK ] 0.53 sec.
2025-04-01 17:41:18 02572_query_views_log_background_thread: [ OK ] 18.18 sec.
2025-04-01 17:41:19 02789_functions_after_sorting_and_columns_with_same_names_bug: [ OK ] 1.03 sec.
2025-04-01 17:41:19 03230_async_insert_native: [ OK ] 4.89 sec.
2025-04-01 17:41:19 02472_segfault_expression_parser: [ OK ] 0.43 sec.
2025-04-01 17:41:20 02481_i43247_ubsan_in_minmaxany: [ OK ] 1.39 sec.
2025-04-01 17:41:21 02718_parquet_metadata_format: [ OK ] 5.01 sec.
2025-04-01 17:41:22 01188_attach_table_from_path: [ OK ] 1.13 sec.
2025-04-01 17:41:23 02715_or_null: [ OK ] 0.58 sec.
2025-04-01 17:41:23 02269_to_start_of_interval_overflow: [ OK ] 0.68 sec.
2025-04-01 17:41:24 02806_cte_block_cannot_be_empty: [ OK ] 0.64 sec.
2025-04-01 17:41:25 02246_flatten_tuple: [ OK ] 0.93 sec.
2025-04-01 17:41:30 02439_merge_selecting_partitions: [ OK ] 10.11 sec.
2025-04-01 17:41:36 01338_long_select_and_alter_zookeeper: [ OK ] 16.06 sec.
2025-04-01 17:41:38 00661_optimize_final_replicated_without_partition_zookeeper: [ OK ] 1.24 sec.
2025-04-01 17:41:38 02029_quantile_sanitizer: [ OK ] 0.73 sec.
2025-04-01 17:41:39 00552_or_nullable: [ OK ] 0.88 sec.
2025-04-01 17:41:41 03271_ghdata_object_to_json_alter: [ OK ] 63.87 sec.
2025-04-01 17:41:41 03174_least_greatest_ignore_null_input_values: [ OK ] 1.33 sec.
2025-04-01 17:41:41 02243_in_ip_address: [ OK ] 0.64 sec.
2025-04-01 17:41:42 03148_setting_max_streams_to_max_threads_ratio_overflow: [ OK ] 1.08 sec.
2025-04-01 17:41:45 01581_deduplicate_by_columns_local: [ OK ] 2.89 sec.
2025-04-01 17:41:45 02346_read_in_order_fixed_prefix: [ OK ] 40.92 sec.
2025-04-01 17:41:45 03271_decimal_monotonic_day_of_week: [ OK ] 0.63 sec.
2025-04-01 17:41:47 02574_suspicious_low_cardinality_msan: [ OK ] 1.28 sec.
2025-04-01 17:41:47 03268_system_parts_index_granularity: [ OK ] 0.54 sec.
2025-04-01 17:41:47 02956_clickhouse_local_system_parts: [ OK ] 2.23 sec.
2025-04-01 17:41:48 01710_query_log_with_projection_info: [ OK ] 6.20 sec.
2025-04-01 17:41:48 01802_toDateTime64_large_values: [ OK ] 0.68 sec.
2025-04-01 17:41:48 00700_decimal_in_keys: [ OK ] 1.08 sec.
2025-04-01 17:41:50 02791_final_block_structure_mismatch_bug: [ OK ] 1.88 sec.
2025-04-01 17:41:50 00411_long_accurate_number_comparison_int1: [ OK ] 64.04 sec.
2025-04-01 17:41:50 02680_default_star: [ OK ] 0.43 sec.
2025-04-01 17:41:50 01825_type_json_ghdata_insert_select: [ OK ] 47.89 sec.
2025-04-01 17:41:50 01423_if_nullable_cond: [ OK ] 0.54 sec.
2025-04-01 17:41:51 01544_file_engine_settings: [ OK ] 2.48 sec.
2025-04-01 17:41:51 02833_std_alias: [ OK ] 0.63 sec.
2025-04-01 17:41:51 02204_fractional_progress_bar_long: [ SKIPPED ] 0.00 sec.
2025-04-01 17:41:51 Reason: not running for current build
2025-04-01 17:41:51 00004_shard_format_ast_and_remote_table: [ OK ] 0.64 sec.
2025-04-01 17:41:51 00804_test_custom_compression_codes_log_storages: [ OK ] 2.33 sec.
2025-04-01 17:41:51 00730_unicode_terminal_format: [ OK ] 0.83 sec.
2025-04-01 17:41:52 02245_s3_virtual_columns: [ OK ] 0.93 sec.
2025-04-01 17:41:52 02566_analyzer_limit_settings_distributed: [ OK ] 0.98 sec.
2025-04-01 17:41:52 03237_max_map_state_decimal_serialization: [ OK ] 0.58 sec.
2025-04-01 17:41:52 00353_join_by_tuple: [ OK ] 0.53 sec.
2025-04-01 17:41:52 01354_tuple_low_cardinality_array_mapped_bug: [ OK ] 0.58 sec.
2025-04-01 17:41:53 02922_server_exit_code: [ OK ] 2.13 sec.
2025-04-01 17:41:54 02417_null_variadic_behaviour: [ OK ] 1.98 sec.
2025-04-01 17:41:55 01834_alias_columns_laziness_filimonov: [ OK ] 3.10 sec.
2025-04-01 17:41:56 01731_async_task_queue_wait: [ OK ] 3.24 sec.
2025-04-01 17:41:56 02097_polygon_dictionary_store_key: [ OK ] 0.98 sec.
2025-04-01 17:41:57 00002_system_numbers: [ OK ] 0.83 sec.
2025-04-01 17:41:57 03008_deduplication_insert_into_partitioned_table: [ OK ] 2.54 sec.
2025-04-01 17:41:57 02982_parallel_replicas_unexpected_cluster: [ OK ] 0.73 sec.
2025-04-01 17:41:58 01070_alter_with_ttl: [ OK ] 0.68 sec.
2025-04-01 17:41:58 02351_Map_combinator_dist: [ OK ] 1.29 sec.
2025-04-01 17:41:59 00520_http_nullable: [ OK ] 1.59 sec.
2025-04-01 17:41:59 00908_bloom_filter_index: [ OK ] 29.45 sec.
2025-04-01 17:41:59 03206_no_exceptions_clickhouse_local: [ FAIL ] 1.78 sec.
2025-04-01 17:41:59 Reason: return code: 134, result:
2025-04-01 17:41:59
2025-04-01 17:41:59
2025-04-01 17:41:59
2025-04-01 17:41:59 stdout:
2025-04-01 17:41:59
2025-04-01 17:41:59
2025-04-01 17:41:59 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 672595 --group_by_two_level_threshold_bytes 22049239 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 1 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 10871429 --max_read_buffer_size 994986 --prefer_localhost_replica 0 --max_block_size 90665 --max_joined_block_size_rows 93016 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 1 --optimize_or_like_chain 1 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 21 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 10920486 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 10737418240 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method mmap --remote_filesystem_read_method read --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 0 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 32Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 19 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 0 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 787984612 --min_compress_block_size 1636648 --max_compress_block_size 2024248 --merge_tree_compact_parts_min_granules_to_multibuffer_read 13 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 8871796 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 3 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 3 --session_timezone Africa/Khartoum --use_page_cache_for_disks_without_file_cache False --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.36 --prefer_external_sort_block_bytes 100000000 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 1 --max_parsing_threads 10 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table false --output_format_native_write_json_as_string 1 --enable_vertical_final 0
2025-04-01 17:41:59
2025-04-01 17:41:59 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 2088974808 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 531140840 --index_granularity_bytes 7269453 --merge_max_block_size 6689 --index_granularity 58306 --min_bytes_for_wide_part 0 --marks_compress_block_size 45718 --primary_key_compress_block_size 81099 --replace_long_file_name_to_hash 0 --max_file_name_length 24 --min_bytes_for_full_part_storage 536870912 --compact_parts_max_bytes_to_buffer 35988816 --compact_parts_max_granules_to_buffer 120 --compact_parts_merge_max_bytes_to_prefetch_part 27544159 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 0 --old_parts_lifetime 10 --prewarm_mark_cache 1 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 1 --use_primary_key_cache 0 --prewarm_primary_key_cache 1
2025-04-01 17:41:59
2025-04-01 17:41:59 Database: test_axigm1xv
2025-04-01 17:42:00 00558_aggregate_merge_totals_with_arenas: [ OK ] 0.58 sec.
2025-04-01 17:42:01 02525_different_engines_in_temporary_tables: [ OK ] 1.33 sec.
2025-04-01 17:42:01 01681_bloom_filter_nullable_column: [ OK ] 1.29 sec.
2025-04-01 17:42:03 00835_if_generic_case: [ OK ] 1.43 sec.
2025-04-01 17:42:03 02423_insert_stats_behaviour: [ OK ] 9.70 sec.
2025-04-01 17:42:03 02538_analyzer_create_table_as_select: [ OK ] 0.68 sec.
2025-04-01 17:42:04 03199_unbin_buffer_overflow: [ OK ] 13.56 sec.
2025-04-01 17:42:04 03003_prql_panic: [ OK ] 1.84 sec.
2025-04-01 17:42:04 01430_fix_any_rewrite_aliases: [ OK ] 0.53 sec.
2025-04-01 17:42:04 00143_number_classification_functions: [ OK ] 0.94 sec.
2025-04-01 17:42:05 03258_old_analyzer_const_expr_bug: [ OK ] 0.53 sec.
2025-04-01 17:42:05 03157_dynamic_type_json: [ OK ] 0.73 sec.
2025-04-01 17:42:05 00752_low_cardinality_lambda_argument: [ OK ] 1.03 sec.
2025-04-01 17:42:06 01825_new_type_json_bools: [ OK ] 0.70 sec.
2025-04-01 17:42:06 01891_partition_by_uuid: [ OK ] 0.64 sec.
2025-04-01 17:42:07 01414_freeze_does_not_prevent_alters: [ OK ] 1.12 sec.
2025-04-01 17:42:07 00506_union_distributed: [ OK ] 1.60 sec.
2025-04-01 17:42:08 01825_new_type_json_multiple_files: [ OK ] 8.45 sec.
2025-04-01 17:42:08 03166_optimize_row_order_during_insert: [ OK ] 1.23 sec.
2025-04-01 17:42:08 00299_stripe_log_multiple_inserts: [ OK ] 1.24 sec.
2025-04-01 17:42:08 00065_shard_float_literals_formatting: [ OK ] 0.68 sec.
2025-04-01 17:42:09 01620_fix_simple_state_arg_type: [ OK ] 0.89 sec.
2025-04-01 17:42:09 02981_variant_type_function: [ OK ] 1.33 sec.
2025-04-01 17:42:09 02811_invalid_embedded_rocksdb_create: [ OK ] 0.65 sec.
2025-04-01 17:42:10 02177_merge_optimize_aggregation_in_order: [ OK ] 0.84 sec.
2025-04-01 17:42:10 02012_compress_lz4: [ OK ] 2.33 sec.
2025-04-01 17:42:11 00809_add_days_segfault: [ OK ] 1.24 sec.
2025-04-01 17:42:11 02882_primary_key_index_in_function_different_types: [ OK ] 0.83 sec.
2025-04-01 17:42:12 00904_array_with_constant_2: [ OK ] 0.58 sec.
2025-04-01 17:42:12 03001_data_version_column: [ OK ] 0.94 sec.
2025-04-01 17:42:12 02477_exists_fuzz_43478: [ OK ] 0.58 sec.
2025-04-01 17:42:12 01493_table_function_null: [ OK ] 0.53 sec.
2025-04-01 17:42:13 00476_pretty_formats_and_widths: [ OK ] 0.63 sec.
2025-04-01 17:42:14 01085_simdjson_uint64: [ OK ] 0.59 sec.
2025-04-01 17:42:14 03032_storage_memory_modify_settings: [ OK ] 1.73 sec.
2025-04-01 17:42:15 03221_merge_profile_events: [ OK ] 14.16 sec.
2025-04-01 17:42:15 02718_cli_dashed_options_parsing: [ OK ] 4.95 sec.
2025-04-01 17:42:15 02752_space_function: [ OK ] 1.58 sec.
2025-04-01 17:42:15 03207_json_read_subcolumns_2_wide_merge_tree: [ SKIPPED ] 0.00 sec.
2025-04-01 17:42:15 Reason: not running for current build
2025-04-01 17:42:16 00974_full_outer_join: [ OK ] 0.70 sec.
2025-04-01 17:42:16 00098_f_union_all: [ OK ] 0.68 sec.
2025-04-01 17:42:17 03036_schema_inference_cache_s3_archives: [ OK ] 0.83 sec.
2025-04-01 17:42:17 03276_index_of_assume_sorted: [ OK ] 0.88 sec.
2025-04-01 17:42:17 01552_dict_fixedstring: [ OK ] 0.73 sec.
2025-04-01 17:42:18 01948_group_bitmap_and_or_xor_fix: [ OK ] 0.68 sec.
2025-04-01 17:42:18 00180_attach_materialized_view: [ OK ] 0.59 sec.
2025-04-01 17:42:18 02236_json_each_row_empty_map_schema_inference: [ OK ] 0.59 sec.
2025-04-01 17:42:19 03272_bad_aggregate_function: [ OK ] 0.73 sec.
2025-04-01 17:42:19 01825_new_type_json_9: [ OK ] 0.68 sec.
2025-04-01 17:42:19 02149_issue_32487: [ OK ] 0.53 sec.
2025-04-01 17:42:20 02680_illegal_type_of_filter_projection: [ OK ] 1.01 sec.
2025-04-01 17:42:21 02003_memory_limit_in_client: [ OK ] 69.11 sec.
2025-04-01 17:42:21 02491_part_log_has_table_uuid: [ OK ] 2.24 sec.
2025-04-01 17:42:22 02293_h3_hex_ring: [ OK ] 1.28 sec.
2025-04-01 17:42:22 03069_analyzer_with_alias_in_array_join: [ OK ] 0.60 sec.
2025-04-01 17:42:23 02474_extract_fixedstring_from_json: [ OK ] 1.28 sec.
2025-04-01 17:42:24 00914_replicate: [ OK ] 0.73 sec.
2025-04-01 17:42:35 02129_skip_quoted_fields: [ OK ] 20.19 sec.
2025-04-01 17:42:37 02562_native_tskv_default_for_omitted_fields: [ OK ] 12.54 sec.
2025-04-01 17:42:38 02346_fulltext_index_search: [ OK ] 40.30 sec.
2025-04-01 17:42:38 00831_quantile_weighted_parameter_check: [ OK ] 1.48 sec.
2025-04-01 17:42:38 00534_functions_bad_arguments8: [ SKIPPED ] 0.00 sec.
2025-04-01 17:42:38 Reason: not running for current build
2025-04-01 17:42:41 01077_mutations_index_consistency: [ OK ] 18.85 sec.
2025-04-01 17:42:41 02504_regexp_dictionary_ua_parser: [ OK ] 26.98 sec.
2025-04-01 17:42:43 01825_type_json_13: [ OK ] 7.05 sec.
2025-04-01 17:42:43 02842_largestTriangleThreeBuckets_aggregate_function: [ OK ] 4.38 sec.
2025-04-01 17:42:43 01034_unknown_qualified_column_in_join: [ OK ] 1.73 sec.
2025-04-01 17:42:46 00712_prewhere_with_final: [ OK ] 2.65 sec.
2025-04-01 17:42:46 02571_local_desc_abort_on_twitter_json: [ OK ] 2.94 sec.
2025-04-01 17:42:47 03009_format_show_database: [ OK ] 3.84 sec.
2025-04-01 17:42:47 00678_shard_funnel_window: [ OK ] 8.85 sec.
2025-04-01 17:42:49 00362_great_circle_distance: [ OK ] 1.40 sec.
2025-04-01 17:42:49 00568_empty_function_with_fixed_string: [ OK ] 1.58 sec.
2025-04-01 17:42:49 01273_arrow_stream: [ OK ] 41.63 sec.
2025-04-01 17:42:51 02841_not_ready_set_bug: [ OK ] 10.09 sec.
2025-04-01 17:42:53 02461_cancel_finish_race: [ OK ] 31.49 sec.
2025-04-01 17:42:53 02027_ngrams: [ OK ] 1.79 sec.
2025-04-01 17:42:53 02355_control_block_size_in_array_join: [ OK ] 3.83 sec.
2025-04-01 17:42:53 01890_state_of_state: [ OK ] 4.12 sec.
2025-04-01 17:42:53 02993_lazy_index_loading: [ OK ] 88.20 sec.
2025-04-01 17:42:54 02325_dates_schema_inference: [ OK ] 1.13 sec.
2025-04-01 17:42:54 01700_system_zookeeper_path_in: [ OK ] 0.83 sec.
2025-04-01 17:42:54 00460_vertical_and_totals_extremes: [ OK ] 0.59 sec.
2025-04-01 17:42:54 01784_parallel_formatting_memory: [ OK ] 1.08 sec.
2025-04-01 17:42:54 02932_query_settings_max_size_drop: [ OK ] 1.08 sec.
2025-04-01 17:42:54 02517_infer_uint64_in_case_of_int64_overflow: [ OK ] 8.70 sec.
2025-04-01 17:42:55 02567_native_type_conversions: [ OK ] 5.20 sec.
2025-04-01 17:42:55 03256_invalid_mutation_query: [ OK ] 1.08 sec.
2025-04-01 17:42:55 01845_add_testcase_for_arrayElement: [ OK ] 0.58 sec.
2025-04-01 17:42:56 01755_shard_pruning_with_literal: [ OK ] 0.63 sec.
2025-04-01 17:42:56 00712_prewhere_with_alias: [ OK ] 1.13 sec.
2025-04-01 17:42:56 01034_JSONCompactEachRow: [ OK ] 1.63 sec.
2025-04-01 17:42:56 01271_show_privileges: [ OK ] 0.53 sec.
2025-04-01 17:42:56 02751_protobuf_ipv6: [ OK ] 2.28 sec.
2025-04-01 17:42:59 02876_s3_cluster_schema_inference_names_with_spaces: [ OK ] 2.64 sec.
2025-04-01 17:42:59 02177_issue_31009: [ SKIPPED ] 0.00 sec.
2025-04-01 17:42:59 Reason: not running for current build
2025-04-01 17:42:59 02165_h3_edge_length_km: [ OK ] 3.00 sec.
2025-04-01 17:43:00 03281_dynamic_coalesce: [ OK ] 3.21 sec.
2025-04-01 17:43:01 01831_max_streams: [ OK ] 1.27 sec.
2025-04-01 17:43:02 01079_bit_operations_using_bitset: [ OK ] 2.45 sec.
2025-04-01 17:43:04 01940_totimezone_operator_monotonicity: [ OK ] 2.75 sec.
2025-04-01 17:43:06 02922_respect_nulls_Nullable: [ OK ] 5.81 sec.
2025-04-01 17:43:07 02479_analyzer_aggregation_crash: [ OK ] 2.00 sec.
2025-04-01 17:43:07 01660_join_or_all: [ OK ] 10.86 sec.
2025-04-01 17:43:08 03267_min_parts_to_merge_at_once: [ OK ] 13.43 sec.
2025-04-01 17:43:08 02883_zookeeper_finalize_stress: [ OK ] 12.56 sec.
2025-04-01 17:43:08 00853_join_with_nulls_crash: [ OK ] 2.08 sec.
2025-04-01 17:43:08 01683_dist_INSERT_block_structure_mismatch: [ OK ] 0.63 sec.
2025-04-01 17:43:09 01893_jit_aggregation_function_min_long: [ OK ] 7.05 sec.
2025-04-01 17:43:09 01079_alter_default_zookeeper_long: [ OK ] 2.19 sec.
2025-04-01 17:43:10 03279_pr_3_way_joins_left_first: [ OK ] 2.48 sec.
2025-04-01 17:43:12 01073_crlf_end_of_line: [ OK ] 1.25 sec.
2025-04-01 17:43:12 03208_array_of_json_read_subcolumns_2_compact_merge_tree: [ SKIPPED ] 0.00 sec.
2025-04-01 17:43:12 Reason: not running for current build
2025-04-01 17:43:13 02457_filesystem_function: [ OK ] 1.27 sec.
2025-04-01 17:43:15 01773_min_max_time_system_parts_datetime64: [ OK ] 1.45 sec.
2025-04-01 17:43:19 03038_recursive_cte_postgres_4: [ OK ] 3.42 sec.
2025-04-01 17:43:21 00938_dataset_test: [ OK ] 1.92 sec.
2025-04-01 17:43:24 03202_dynamic_null_map_subcolumn: [ OK ] 14.29 sec.
2025-04-01 17:43:25 00742_require_join_strictness: [ OK ] 3.74 sec.
2025-04-01 17:43:28 00975_json_hang: [ OK ] 19.26 sec.
2025-04-01 17:43:31 03199_queries_with_new_analyzer: [ OK ] 4.90 sec.
2025-04-01 17:43:32 02504_disallow_arrayjoin_in_mutations: [ OK ] 3.40 sec.
2025-04-01 17:43:32 03129_cte_with_final: [ OK ] 1.43 sec.
2025-04-01 17:43:33 00612_union_query_with_subquery: [ OK ] 1.07 sec.
2025-04-01 17:43:36 01463_resample_overflow: [ OK ] 2.96 sec.
2025-04-01 17:43:37 02125_lz4_compression_bug_JSONCompactEachRow: [ OK ] 42.98 sec.
2025-04-01 17:43:38 02163_operators: [ OK ] 1.27 sec.
2025-04-01 17:43:41 02255_broken_parts_chain_on_start: [ OK ] 15.26 sec.
2025-04-01 17:43:41 02366_kql_create_table: [ OK ] 2.71 sec.
2025-04-01 17:43:42 00503_cast_const_nullable: [ OK ] 0.59 sec.
2025-04-01 17:43:42 00991_system_parts_race_condition_long: [ OK ] 34.02 sec.
2025-04-01 17:43:42 02931_alter_materialized_view_query_inconsistent: [ OK ] 1.49 sec.
2025-04-01 17:43:43 01404_roundUpToPowerOfTwoOrZero_safety: [ OK ] 0.68 sec.
2025-04-01 17:43:43 01072_optimize_skip_unused_shards_const_expr_eval: [ OK ] 7.42 sec.
2025-04-01 17:43:43 01060_window_view_event_tumble_to_asc: [ OK ] 10.03 sec.
2025-04-01 17:43:44 00411_merge_tree_where_const_in_set: [ OK ] 0.78 sec.
2025-04-01 17:43:44 01753_mutate_table_predicated_with_table: [ OK ] 0.73 sec.
2025-04-01 17:43:45 02494_array_function_range: [ OK ] 0.68 sec.
2025-04-01 17:43:45 02176_toStartOfWeek_overflow_pruning: [ OK ] 0.83 sec.
2025-04-01 17:43:45 01621_clickhouse_compressor: [ OK ] 2.03 sec.
2025-04-01 17:43:45 01326_hostname_alias: [ OK ] 0.53 sec.
2025-04-01 17:43:46 01412_row_from_totals: [ OK ] 0.94 sec.
2025-04-01 17:43:47 01268_mv_scalars: [ OK ] 1.41 sec.
2025-04-01 17:43:48 01359_codeql: [ OK ] 0.48 sec.
2025-04-01 17:43:48 00814_parsing_ub: [ OK ] 0.53 sec.
2025-04-01 17:43:49 03246_json_simd_rapid_parsers: [ OK ] 3.44 sec.
2025-04-01 17:43:49 02815_range_dict_no_direct_join: [ OK ] 0.93 sec.
2025-04-01 17:43:50 03156_nullable_number_tips: [ OK ] 0.98 sec.
2025-04-01 17:43:50 02263_format_insert_settings: [ OK ] 8.20 sec.
2025-04-01 17:43:50 00358_from_string_complex_types: [ OK ] 0.53 sec.
2025-04-01 17:43:50 01460_line_as_string_format: [ OK ] 63.65 sec.
2025-04-01 17:43:51 02387_parse_date_as_datetime: [ OK ] 0.63 sec.
2025-04-01 17:43:51 01049_join_low_card_crash: [ OK ] 0.93 sec.
2025-04-01 17:43:51 01019_array_fill: [ OK ] 0.83 sec.
2025-04-01 17:43:51 00686_client_exit_code: [ OK ] 1.74 sec.
2025-04-01 17:43:52 00975_sample_prewhere_distributed: [ OK ] 0.73 sec.
2025-04-01 17:43:52 00296_url_parameters: [ OK ] 0.93 sec.
2025-04-01 17:43:53 01056_predicate_optimizer_bugs: [ OK ] 1.93 sec.
2025-04-01 17:43:54 02016_aggregation_spark_bar: [ OK ] 2.28 sec.
2025-04-01 17:43:55 02454_create_table_with_custom_disk: [ OK ] 1.14 sec.
2025-04-01 17:43:56 03234_check_grant: [ OK ] 4.59 sec.
2025-04-01 17:43:56 00623_truncate_table: [ OK ] 1.99 sec.
2025-04-01 17:43:57 03161_cnf_reduction: [ OK ] 1.73 sec.
2025-04-01 17:43:57 03150_infer_type_variant: [ OK ] 0.58 sec.
2025-04-01 17:43:58 00662_has_nullable: [ OK ] 0.83 sec.
2025-04-01 17:43:59 02122_parallel_formatting_TSV: [ OK ] 7.15 sec.
2025-04-01 17:44:00 01547_query_log_current_database: [ OK ] 4.29 sec.
2025-04-01 17:44:01 00417_kill_query: [ OK ] 5.30 sec.
2025-04-01 17:44:03 00926_adaptive_index_granularity_collapsing_merge_tree: [ OK ] 1.54 sec.
2025-04-01 17:44:04 00842_array_with_constant_overflow: [ OK ] 0.99 sec.
2025-04-01 17:44:06 02661_read_from_archive_tzst: [ OK ] 20.48 sec.
2025-04-01 17:44:06 01505_log_distributed_deadlock: [ OK ] 0.69 sec.
2025-04-01 17:44:07 01144_multiword_data_types: [ OK ] 0.79 sec.
2025-04-01 17:44:08 00351_select_distinct_arrays_tuples: [ OK ] 0.64 sec.
2025-04-01 17:44:09 03113_analyzer_not_found_column_in_block_2: [ OK ] 0.78 sec.
2025-04-01 17:44:10 03130_analyzer_self_join_group_by: [ OK ] 1.03 sec.
2025-04-01 17:44:10 01632_tinylog_read_write: [ OK ] 11.82 sec.
2025-04-01 17:44:11 02732_rename_after_processing: [ OK ] 6.65 sec.
2025-04-01 17:44:12 01702_rewrite_avg_for_algebraic_optimization: [ OK ] 1.03 sec.
2025-04-01 17:44:13 01272_suspicious_codecs: [ OK ] 2.99 sec.
2025-04-01 17:44:14 00338_replicate_array_of_strings: [ OK ] 1.04 sec.
2025-04-01 17:44:15 02271_replace_partition_many_tables: [ OK ] 31.51 sec.
2025-04-01 17:44:15 02235_add_part_offset_virtual_column: [ OK ] 15.67 sec.
2025-04-01 17:44:15 01440_big_int_shift: [ OK ] 0.68 sec.
2025-04-01 17:44:15 01640_distributed_async_insert_compression: [ OK ] 0.79 sec.
2025-04-01 17:44:16 00465_nullable_default: [ OK ] 0.65 sec.
2025-04-01 17:44:16 01837_cast_to_array_from_empty_array: [ OK ] 0.59 sec.
2025-04-01 17:44:16 02394_every_profile_event_must_have_documentation: [ OK ] 0.55 sec.
2025-04-01 17:44:17 00803_odbc_driver_2_format: [ OK ] 0.50 sec.
2025-04-01 17:44:17 00910_buffer_prewhere: [ OK ] 0.67 sec.
2025-04-01 17:44:18 01825_type_json_3: [ OK ] 2.79 sec.
2025-04-01 17:44:18 01276_system_licenses: [ OK ] 1.15 sec.
2025-04-01 17:44:18 00165_transform_non_const_default: [ OK ] 1.00 sec.
2025-04-01 17:44:19 01011_group_uniq_array_memsan: [ OK ] 0.61 sec.
2025-04-01 17:44:19 03013_position_const_start_pos: [ OK ] 0.53 sec.
2025-04-01 17:44:19 01720_engine_file_empty_if_not_exists: [ OK ] 1.55 sec.
2025-04-01 17:44:20 01038_dictionary_lifetime_min_zero_sec: [ OK ] 7.87 sec.
2025-04-01 17:44:20 01851_s2_to_geo: [ OK ] 0.59 sec.
2025-04-01 17:44:20 02902_select_subcolumns_from_engine_null: [ OK ] 0.65 sec.
2025-04-01 17:44:20 02206_clickhouse_local_use_database: [ OK ] 1.94 sec.
2025-04-01 17:44:21 02923_join_use_nulls_modulo: [ OK ] 0.95 sec.
2025-04-01 17:44:22 01790_dist_INSERT_block_structure_mismatch_types_and_names: [ OK ] 1.44 sec.
2025-04-01 17:44:23 02317_functions_with_nothing: [ OK ] 1.39 sec.
2025-04-01 17:44:24 01849_geoToS2: [ OK ] 2.36 sec.
2025-04-01 17:44:24 02475_or_function_alias_and_const_where: [ OK ] 0.63 sec.
2025-04-01 17:44:25 02967_fuzz_bad_cast: [ OK ] 1.19 sec.
2025-04-01 17:44:26 03084_analyzer_join_column_alias: [ OK ] 1.59 sec.
2025-04-01 17:44:26 01498_alter_column_storage_memory: [ OK ] 0.71 sec.
2025-04-01 17:44:26 00263_merge_aggregates_and_overflow: [ OK ] 1.15 sec.
2025-04-01 17:44:29 02813_seriesOutliersDetectTukey: [ OK ] 2.41 sec.
2025-04-01 17:44:29 01064_arrayROCAUC: [ OK ] 2.71 sec.
2025-04-01 17:44:31 00688_low_cardinality_syntax: [ OK ] 1.80 sec.
2025-04-01 17:44:31 01772_to_start_of_hour_align: [ OK ] 1.50 sec.
2025-04-01 17:44:32 02540_analyzer_matcher_alias_materialized_columns: [ OK ] 0.95 sec.
2025-04-01 17:44:32 00521_multidimensional: [ OK ] 1.39 sec.
2025-04-01 17:44:34 02990_format_select_from_explain: [ OK ] 1.43 sec.
2025-04-01 17:44:34 03273_group_by_in_order_still_used_when_group_by_key_doesnt_match_order_by_key: [ OK ] 2.04 sec.
2025-04-01 17:44:35 00751_low_cardinality_nullable_group_by: [ OK ] 34.38 sec.
2025-04-01 17:44:35 02245_s3_schema_desc: [ OK ] 1.18 sec.
2025-04-01 17:44:35 02933_group_by_memory_usage: [ OK ] 25.74 sec.
2025-04-01 17:44:36 02915_fpc_overflow: [ OK ] 1.84 sec.
2025-04-01 17:44:36 00950_test_gorilla_codec: [ OK ] 1.18 sec.
2025-04-01 17:44:36 01925_json_as_string_data_in_square_brackets: [ OK ] 0.64 sec.
2025-04-01 17:44:36 03227_dynamic_subcolumns_enumerate_streams: [ OK ] 0.63 sec.
2025-04-01 17:44:37 02354_with_statement_non_exist_column: [ OK ] 0.60 sec.
2025-04-01 17:44:37 02862_sorted_distinct_sparse_fix: [ OK ] 0.81 sec.
2025-04-01 17:44:37 01088_array_slice_of_aggregate_functions: [ OK ] 0.65 sec.
2025-04-01 17:44:37 02877_optimize_read_in_order_from_view: [ OK ] 17.49 sec.
2025-04-01 17:44:37 02455_extract_fixed_string_from_nested_json: [ OK ] 0.59 sec.
2025-04-01 17:44:37 01651_group_uniq_array_enum: [ OK ] 0.58 sec.
2025-04-01 17:44:38 02265_per_table_ttl_mutation_on_change: [ OK ] 1.39 sec.
2025-04-01 17:44:38 02963_single_value_destructor: [ OK ] 1.13 sec.
2025-04-01 17:44:38 01009_insert_select_data_loss: [ OK ] 0.63 sec.
2025-04-01 17:44:38 02476_fix_cast_parser_bug: [ OK ] 0.43 sec.
2025-04-01 17:44:38 00950_bad_alloc_when_truncate_join_storage: [ OK ] 0.48 sec.
2025-04-01 17:44:39 01259_datetime64_ubsan: [ OK ] 0.84 sec.
2025-04-01 17:44:39 02266_auto_add_nullable: [ OK ] 0.63 sec.
2025-04-01 17:44:39 01451_wrong_error_long_query: [ OK ] 1.78 sec.
2025-04-01 17:44:39 00284_external_aggregation_2: [ OK ] 91.78 sec.
2025-04-01 17:44:39 00609_distributed_with_case_when_then: [ OK ] 0.73 sec.
2025-04-01 17:44:39 03276_functions_to_subcolumns_lc: [ OK ] 0.63 sec.
2025-04-01 17:44:40 01906_bigint_accurate_cast_ubsan: [ OK ] 1.43 sec.
2025-04-01 17:44:41 02674_trivial_count_analyzer: [ OK ] 1.43 sec.
2025-04-01 17:44:41 02874_toDaysSinceYearZero: [ OK ] 1.92 sec.
2025-04-01 17:44:42 02002_system_table_with_tuple: [ OK ] 2.00 sec.
2025-04-01 17:44:42 00628_in_lambda_on_merge_table_bug: [ OK ] 0.88 sec.
2025-04-01 17:44:43 01714_alter_drop_version: [ OK ] 0.98 sec.
2025-04-01 17:44:44 01461_query_start_time_microseconds: [ OK ] 6.75 sec.
2025-04-01 17:44:44 02950_parallel_replicas_used_count: [ OK ] 24.20 sec.
2025-04-01 17:44:44 02943_variant_element: [ OK ] 0.78 sec.
2025-04-01 17:44:45 03043_group_array_result_is_expected: [ OK ] 0.63 sec.
2025-04-01 17:44:45 01690_quantilesTiming_ubsan: [ OK ] 0.48 sec.
2025-04-01 17:44:45 01081_demangle: [ OK ] 0.53 sec.
2025-04-01 17:44:45 00287_column_const_with_nan: [ OK ] 0.52 sec.
2025-04-01 17:44:45 01392_column_resolve: [ OK ] 0.68 sec.
2025-04-01 17:44:48 01825_new_type_json_11: [ OK ] 6.80 sec.
2025-04-01 17:44:49 00590_limit_by_column_removal: [ OK ] 0.58 sec.
2025-04-01 17:44:50 02908_Npy_files_caching: [ OK ] 4.69 sec.
2025-04-01 17:44:51 02276_full_sort_join_unsupported: [ OK ] 1.63 sec.
2025-04-01 17:44:51 01014_count_of_merges_metrics: [ OK ] 0.74 sec.
2025-04-01 17:44:51 02481_default_value_used_in_row_level_filter: [ OK ] 0.73 sec.
2025-04-01 17:44:52 02481_low_cardinality_with_short_circuit_functins: [ OK ] 0.83 sec.
2025-04-01 17:44:52 02946_literal_alias_misclassification: [ OK ] 0.63 sec.
2025-04-01 17:44:53 01055_prewhere_bugs: [ OK ] 0.83 sec.
2025-04-01 17:44:54 01646_rewrite_sum_if: [ OK ] 1.23 sec.
2025-04-01 17:44:57 02461_prewhere_row_level_policy_lightweight_delete: [ OK ] 4.94 sec.
2025-04-01 17:44:57 00844_join_lightee2: [ OK ] 0.78 sec.
2025-04-01 17:44:59 02225_hints_for_indeices: [ OK ] 4.39 sec.
2025-04-01 17:44:59 00647_select_numbers_with_offset: [ OK ] 0.53 sec.
2025-04-01 17:45:00 02971_limit_by_distributed: [ OK ] 0.83 sec.
2025-04-01 17:45:02 01520_client_print_query_id: [ OK ] 1.93 sec.
2025-04-01 17:45:03 00964_bloom_index_string_functions: [ OK ] 17.37 sec.
2025-04-01 17:45:03 00757_enum_defaults_const: [ OK ] 0.64 sec.
2025-04-01 17:45:04 01013_totals_without_aggregation: [ OK ] 0.89 sec.
2025-04-01 17:45:04 00875_join_right_nulls_ors: [ OK ] 1.03 sec.
2025-04-01 17:45:05 01880_remote_ipv6: [ OK ] 1.13 sec.
2025-04-01 17:45:05 01910_view_dictionary_check_refresh: [ OK ] 25.19 sec.
2025-04-01 17:45:05 02941_variant_type_2: [ OK ] 118.67 sec.
2025-04-01 17:45:05 02962_analyzer_constant_set: [ OK ] 0.58 sec.
2025-04-01 17:45:06 02366_kql_distinct: [ OK ] 0.73 sec.
2025-04-01 17:45:06 03007_column_nullable_uninitialzed_value: [ OK ] 0.53 sec.
2025-04-01 17:45:06 02711_trim_aliases: [ OK ] 0.63 sec.
2025-04-01 17:45:07 00668_compare_arrays_silviucpp: [ OK ] 0.58 sec.
2025-04-01 17:45:07 01328_bad_peephole_optimization: [ OK ] 0.53 sec.
2025-04-01 17:45:07 02711_server_uuid_macro: [ OK ] 1.03 sec.
2025-04-01 17:45:08 02337_join_analyze_stuck: [ OK ] 2.18 sec.
2025-04-01 17:45:08 02504_parse_datetime_best_effort_calebeaires: [ OK ] 0.58 sec.
2025-04-01 17:45:08 00072_in_types: [ OK ] 0.53 sec.
2025-04-01 17:45:09 01093_cyclic_defaults_filimonov: [ OK ] 0.68 sec.
2025-04-01 17:45:09 02931_size_virtual_column_use_structure_from_insertion_table: [ OK ] 1.83 sec.
2025-04-01 17:45:09 02418_keeper_map_keys_limit: [ OK ] 1.09 sec.
2025-04-01 17:45:10 00152_totals_in_subquery: [ OK ] 0.54 sec.
2025-04-01 17:45:10 01037_zookeeper_check_table_empty_pk: [ OK ] 0.83 sec.
2025-04-01 17:45:11 02243_make_date32: [ OK ] 1.98 sec.
2025-04-01 17:45:11 02999_analyzer_preimage_null: [ OK ] 0.63 sec.
2025-04-01 17:45:11 00340_squashing_insert_select: [ OK ] 32.61 sec.
2025-04-01 17:45:11 03001_bad_error_message_higher_order_functions: [ OK ] 2.08 sec.
2025-04-01 17:45:12 03199_fix_auc_tie_handling: [ OK ] 0.68 sec.
2025-04-01 17:45:13 02895_peak_memory_usage_http_headers_regression: [ OK ] 2.03 sec.
2025-04-01 17:45:14 01788_update_nested_type_subcolumn_check: [ OK ] 2.09 sec.
2025-04-01 17:45:14 02301_harmful_reexec: [ OK ] 2.33 sec.
2025-04-01 17:45:14 03156_group_concat: [ OK ] 2.68 sec.
2025-04-01 17:45:14 02049_lowcardinality_shortcircuit_crash: [ OK ] 0.68 sec.
2025-04-01 17:45:15 00802_daylight_saving_time_shift_backwards_at_midnight: [ OK ] 0.63 sec.
2025-04-01 17:45:15 01120_join_constants: [ OK ] 0.68 sec.
2025-04-01 17:45:16 01773_datetime64_add_ubsan: [ OK ] 0.78 sec.
2025-04-01 17:45:16 01821_join_table_mutation: [ OK ] 0.98 sec.
2025-04-01 17:45:16 01747_alter_partition_key_enum_zookeeper_long: [ OK ] 1.53 sec.
2025-04-01 17:45:16 01102_distributed_local_in_bug: [ OK ] 0.78 sec.
2025-04-01 17:45:17 01937_nested_chinese: [ OK ] 0.68 sec.
2025-04-01 17:45:17 03145_unicode_quotes: [ OK ] 0.58 sec.
2025-04-01 17:45:17 01122_totals_rollup_having_block_header: [ OK ] 0.58 sec.
2025-04-01 17:45:17 03033_final_undefined_last_mark: [ OK ] 0.44 sec.
2025-04-01 17:45:18 02150_replace_regexp_all_empty_match: [ OK ] 0.48 sec.
2025-04-01 17:45:18 00732_quorum_insert_simple_test_2_parts_zookeeper_long: [ OK ] 0.98 sec.
2025-04-01 17:45:18 02346_fulltext_index_bug47393: [ OK ] 0.78 sec.
2025-04-01 17:45:19 03233_dynamic_in_functions: [ OK ] 5.84 sec.
2025-04-01 17:45:19 00169_join_constant_keys: [ OK ] 0.95 sec.
2025-04-01 17:45:19 00534_functions_bad_arguments2: [ SKIPPED ] 0.00 sec.
2025-04-01 17:45:19 Reason: not running for current build
2025-04-01 17:45:20 02493_analyzer_sum_if_to_count_if: [ OK ] 1.13 sec.
2025-04-01 17:45:20 01720_union_distinct_with_limit: [ OK ] 0.58 sec.
2025-04-01 17:45:21 01419_merge_tree_settings_sanity_check: [ OK ] 1.08 sec.
2025-04-01 17:45:21 00217_shard_global_subquery_columns_with_same_name: [ OK ] 0.63 sec.
2025-04-01 17:45:21 03234_proto_complex_nested_repeated_noexception: [ OK ] 2.22 sec.
2025-04-01 17:45:22 03006_buffer_overflow_join: [ OK ] 0.58 sec.
2025-04-01 17:45:22 02985_minmax_index_aggregate_function: [ OK ] 0.98 sec.
2025-04-01 17:45:23 02591_bson_long_tuple: [ OK ] 0.58 sec.
2025-04-01 17:45:23 02473_multistep_split_prewhere: [ OK ] 38.07 sec.
2025-04-01 17:45:24 02764_index_analysis_fix: [ OK ] 0.63 sec.
2025-04-01 17:45:24 02226_low_cardinality_text_bloom_filter_index: [ OK ] 1.68 sec.
2025-04-01 17:45:25 01889_check_row_policy_defined_using_user_function: [ OK ] 9.16 sec.
2025-04-01 17:45:25 02112_delayed_clickhouse_client_with_queries_file: [ OK ] 2.08 sec.
2025-04-01 17:45:25 00758_array_reverse: [ OK ] 0.99 sec.
2025-04-01 17:45:26 01548_create_table_compound_column_format: [ OK ] 1.69 sec.
2025-04-01 17:45:28 00972_geohashesInBox: [ OK ] 3.94 sec.
2025-04-01 17:45:28 02128_apply_lambda_parsing: [ OK ] 0.59 sec.
2025-04-01 17:45:28 02461_alter_update_respect_part_column_type_bug: [ OK ] 3.64 sec.
2025-04-01 17:45:29 03040_dynamic_type_alters_1_wide_merge_tree: [ OK ] 2.70 sec.
2025-04-01 17:45:29 02810_fix_remove_dedundant_distinct_view: [ OK ] 0.94 sec.
2025-04-01 17:45:30 01892_setting_limit_offset_distributed: [ OK ] 0.99 sec.
2025-04-01 17:45:31 02150_index_hypothesis_race_long: [ OK ] 10.80 sec.
2025-04-01 17:45:31 01442_merge_detach_attach_long: [ OK ] 33.91 sec.
2025-04-01 17:45:32 02835_fuzz_remove_redundant_sorting: [ OK ] 2.24 sec.
2025-04-01 17:45:32 02884_duplicate_index_name: [ OK ] 0.58 sec.
2025-04-01 17:45:33 02316_cast_to_ip_address_default_column: [ OK ] 0.98 sec.
2025-04-01 17:45:33 01451_replicated_detach_drop_part_long: [ OK ] 2.04 sec.
2025-04-01 17:45:34 01280_unicode_whitespaces_lexer: [ OK ] 0.59 sec.
2025-04-01 17:45:35 02041_test_fuzzy_alter: [ OK ] 0.74 sec.
2025-04-01 17:45:35 01710_projection_optimize_group_by_function_keys: [ OK ] 0.68 sec.
2025-04-01 17:45:37 00100_subquery_table_identifier: [ OK ] 4.19 sec.
2025-04-01 17:45:42 01099_operators_date_and_timestamp: [ OK ] 6.09 sec.
2025-04-01 17:45:42 00953_constraints_operations: [ OK ] 10.50 sec.
2025-04-01 17:45:43 01580_column_const_comparision: [ OK ] 0.63 sec.
2025-04-01 17:45:43 01415_sticking_mutations: [ OK ] 39.42 sec.
2025-04-01 17:45:44 01021_only_tuple_columns: [ OK ] 2.24 sec.
2025-04-01 17:45:44 00726_length_aliases: [ OK ] 0.55 sec.
2025-04-01 17:45:45 03230_date_trunc_and_to_start_of_interval_on_date32: [ OK ] 1.50 sec.
2025-04-01 17:45:45 01199_url_functions_path_without_schema_yiurule: [ OK ] 0.58 sec.
2025-04-01 17:45:46 01880_materialized_view_to_table_type_check: [ OK ] 1.08 sec.
2025-04-01 17:45:46 03200_subcolumns_join_use_nulls: [ OK ] 0.79 sec.
2025-04-01 17:45:47 03152_analyzer_columns_list: [ OK ] 0.98 sec.
2025-04-01 17:45:47 01550_mutation_subquery: [ OK ] 0.68 sec.
2025-04-01 17:45:47 02477_age: [ OK ] 1.53 sec.
2025-04-01 17:45:48 03031_clickhouse_local_input: [ OK ] 4.64 sec.
2025-04-01 17:45:48 01085_extract_all_empty: [ OK ] 0.53 sec.
2025-04-01 17:45:49 00357_to_string_complex_types: [ OK ] 0.78 sec.
2025-04-01 17:45:49 03215_validate_type_in_alter_add_modify_column: [ OK ] 1.58 sec.
2025-04-01 17:45:50 00952_part_frozen_info: [ OK ] 0.89 sec.
2025-04-01 17:45:50 03062_analyzer_join_engine_missing_column: [ OK ] 0.68 sec.
2025-04-01 17:45:50 01663_quantile_weighted_overflow: [ OK ] 0.58 sec.
2025-04-01 17:45:51 03266_with_fill_staleness: [ OK ] 0.88 sec.
2025-04-01 17:45:51 01398_in_tuple_func: [ OK ] 0.69 sec.
2025-04-01 17:45:51 01710_projections_optimize_aggregation_in_order: [ OK ] 21.13 sec.
2025-04-01 17:45:52 00055_join_two_numbers: [ OK ] 0.63 sec.
2025-04-01 17:45:52 02267_insert_empty_data: [ OK ] 0.53 sec.
2025-04-01 17:45:53 03140_client_subsequent_external_tables: [ OK ] 1.93 sec.
2025-04-01 17:46:03 02841_join_filter_set_sparse: [ FAIL ] 1.08 sec.
2025-04-01 17:46:03 Reason: return code: 32
2025-04-01 17:46:03 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-01 17:46:03 (query: SELECT countIf(ignore(*) == 0) FROM t1 JOIN t2 ON t1.s = t2.s;)
2025-04-01 17:46:03 , result:
2025-04-01 17:46:03
2025-04-01 17:46:03
2025-04-01 17:46:03
2025-04-01 17:46:03 stdout:
2025-04-01 17:46:03
2025-04-01 17:46:03
2025-04-01 17:46:03 Settings used in the test: --max_insert_threads 3 --group_by_two_level_threshold 482426 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 0 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 21376022 --max_read_buffer_size 635170 --prefer_localhost_replica 0 --max_block_size 75845 --max_joined_block_size_rows 54219 --max_threads 2 --optimize_append_index 0 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 0 --optimize_or_like_chain 0 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 91 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 44115711 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 1 --min_bytes_to_use_mmap_io 1 --local_filesystem_read_method read --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 10 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 64Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 1Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 0 --merge_tree_coarse_index_granularity 22 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 0 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 988381737 --min_compress_block_size 1817107 --max_compress_block_size 284446 --merge_tree_compact_parts_min_granules_to_multibuffer_read 49 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 7804188 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Africa/Juba --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.3 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 1 --max_parsing_threads 1 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 0 --query_plan_join_swap_table true --output_format_native_write_json_as_string 1 --enable_vertical_final 0
2025-04-01 17:46:03
2025-04-01 17:46:03 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 6240349003 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 100 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 3664758498 --index_granularity_bytes 330795 --merge_max_block_size 19699 --index_granularity 11246 --min_bytes_for_wide_part 1073741824 --marks_compress_block_size 35030 --primary_key_compress_block_size 34332 --replace_long_file_name_to_hash 0 --max_file_name_length 55 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 510861828 --compact_parts_max_granules_to_buffer 175 --compact_parts_merge_max_bytes_to_prefetch_part 12289243 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 35 --old_parts_lifetime 480 --prewarm_mark_cache 0 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 1 --use_primary_key_cache 1 --prewarm_primary_key_cache 0
2025-04-01 17:46:03
2025-04-01 17:46:03 Database: test_rnvzzqsk
2025-04-01 17:46:03 01671_aggregate_function_group_bitmap_data: [ FAIL ] 0.68 sec.
2025-04-01 17:46:03 Reason: return code: 210
2025-04-01 17:46:03 Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:58584): (localhost:9000, ::1, local address: [::1]:58584). (NETWORK_ERROR)
2025-04-01 17:46:03
2025-04-01 17:46:03 , result:
2025-04-01 17:46:03
2025-04-01 17:46:03
2025-04-01 17:46:03
2025-04-01 17:46:03 stdout:
2025-04-01 17:46:03
2025-04-01 17:46:03
2025-04-01 17:46:03 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 1 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 1 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 12201714 --max_read_buffer_size 865507 --prefer_localhost_replica 0 --max_block_size 97703 --max_joined_block_size_rows 29130 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 0 --optimize_or_like_chain 0 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 32 --optimize_aggregation_in_order 0 --aggregation_in_order_max_block_bytes 34195147 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 6725307069 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method io_uring --remote_filesystem_read_method read --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 1 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 64Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 8Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 1 --compile_sort_description 1 --merge_tree_coarse_index_granularity 22 --optimize_distinct_in_order 0 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 337055548 --min_compress_block_size 2434648 --max_compress_block_size 153032 --merge_tree_compact_parts_min_granules_to_multibuffer_read 8 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 1300899 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 0 --session_timezone Africa/Juba --use_page_cache_for_disks_without_file_cache False --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.57 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 1 --max_parsing_threads 10 --optimize_functions_to_subcolumns 1 --parallel_replicas_local_plan 0 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 0 --enable_vertical_final 0
2025-04-01 17:46:03
2025-04-01 17:46:03 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 1 --vertical_merge_algorithm_min_rows_to_activate 1 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 5894704 --merge_max_block_size 3456 --index_granularity 6361 --min_bytes_for_wide_part 0 --marks_compress_block_size 38760 --primary_key_compress_block_size 61354 --replace_long_file_name_to_hash 1 --max_file_name_length 46 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 28684583 --compact_parts_max_granules_to_buffer 253 --compact_parts_merge_max_bytes_to_prefetch_part 4049434 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 100 --old_parts_lifetime 480 --prewarm_mark_cache 1 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 0 --use_primary_key_cache 0 --prewarm_primary_key_cache 0
2025-04-01 17:46:03
2025-04-01 17:46:03 Database: test_umfejymz
2025-04-01 17:46:03 01681_hyperscan_debug_assertion: [ FAIL ] 11.92 sec.
2025-04-01 17:46:03 Reason: server died
2025-04-01 17:46:03 ConnectionRefusedError
2025-04-01 17:46:03 [Errno 111] Connection refused
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 1854, in run
2025-04-01 17:46:03 self._cleanup(result.status == TestStatus.OK)
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 1917, in _cleanup
2025-04-01 17:46:03 self._cleanup_database(args, timeout)
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 1951, in _cleanup_database
2025-04-01 17:46:03 clickhouse_execute(
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 334, in clickhouse_execute
2025-04-01 17:46:03 return clickhouse_execute_http(
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 315, in clickhouse_execute_http
2025-04-01 17:46:03 raise ex
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 304, in clickhouse_execute_http
2025-04-01 17:46:03 client.request(
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/lib/python3.10/http/client.py", line 1283, in request
2025-04-01 17:46:03 self._send_request(method, url, body, headers, encode_chunked)
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/lib/python3.10/http/client.py", line 1329, in _send_request
2025-04-01 17:46:03 self.endheaders(body, encode_chunked=encode_chunked)
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/lib/python3.10/http/client.py", line 1278, in endheaders
2025-04-01 17:46:03 self._send_output(message_body, encode_chunked=encode_chunked)
2025-04-01 17:46:03
2025-04-01 17:46:03 File "/usr/lib/python3.10/http/client.py", line 1038, in _send_output
2025-04-01 17:46:03 self.send(msg)
2025-04-01 17:46:03
2025-04-01 17:46:03 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 1000000 --group_by_two_level_threshold_bytes 8003255 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 1 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 12313890 --max_read_buffer_size 719803 --prefer_localhost_replica 0 --max_block_size 52379 --max_joined_block_size_rows 66586 --max_threads 3 --optimize_append_index 0 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 0 --optimize_or_like_chain 1 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 39 --optimize_aggregation_in_order 0 --aggregation_in_order_max_block_bytes 17533293 --use_uncompressed_cache 1 --min_bytes_to_use_direct_io 10737418240 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method mmap --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 10 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 128Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 1Mi --filesystem_prefetch_step_marks 50 --filesystem_prefetch_step_bytes 100Mi --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 20 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 0 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 694971537 --min_compress_block_size 2979773 --max_compress_block_size 2457989 --merge_tree_compact_parts_min_granules_to_multibuffer_read 55 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 9866312 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.83 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 0 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 0 --max_parsing_threads 0 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table true --output_format_native_write_json_as_string 0 --enable_vertical_final 0
2025-04-01 17:46:03
2025-04-01 17:46:03 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 3998079087 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 100 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 1 --index_granularity_bytes 31337897 --merge_max_block_size 3528 --index_granularity 42523 --min_bytes_for_wide_part 273598252 --marks_compress_block_size 50847 --primary_key_compress_block_size 47749 --replace_long_file_name_to_hash 1 --max_file_name_length 128 --min_bytes_for_full_part_storage 55418150 --compact_parts_max_bytes_to_buffer 450913845 --compact_parts_max_granules_to_buffer 168 --compact_parts_merge_max_bytes_to_prefetch_part 7133480 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 32 --old_parts_lifetime 10 --prewarm_mark_cache 0 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 1 --use_primary_key_cache 1 --prewarm_primary_key_cache 0
2025-04-01 17:46:03
2025-04-01 17:46:03 Database: test_ofjicghs
2025-04-01 17:46:03 Process Process-9:
2025-04-01 17:46:03 Traceback (most recent call last):
2025-04-01 17:46:03 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-01 17:46:03 self.run()
2025-04-01 17:46:03 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-01 17:46:03 self._target(*self._args, **self._kwargs)
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-01 17:46:03 return run_tests_array(*args, **kwargs)
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-01 17:46:03 stop_tests()
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:46:03 cleanup_child_processes(os.getpid())
2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:46:03 child_pgid = os.getpgid(child)
2025-04-01 17:46:03 ProcessLookupError: [Errno 3] No such process
2025-04-01 17:46:03 Child processes of 1396:
2025-04-01 17:46:03 41785 /bin/sh -c pgrep --parent 1396 -a
2025-04-01 17:46:04 02844_max_backup_bandwidth_s3: [ FAIL ] 6.00 sec.
2025-04-01 17:46:04 Reason: return code: 210
2025-04-01 17:46:04 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-01 17:46:04 (query: SYSTEM FLUSH LOGS;)
2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:04
2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:04
2025-04-01 17:46:04 , result:
2025-04-01 17:46:04
2025-04-01 17:46:04
2025-04-01 17:46:04
2025-04-01 17:46:04 stdout:
2025-04-01 17:46:04
2025-04-01 17:46:04
2025-04-01 17:46:04 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 741683 --group_by_two_level_threshold_bytes 47159360 --distributed_aggregation_memory_efficient 1 --fsync_metadata 1 --output_format_parallel_formatting 0 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 12317057 --max_read_buffer_size 754288 --prefer_localhost_replica 1 --max_block_size 36740 --max_joined_block_size_rows 98736 --max_threads 1 --optimize_append_index 1 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 1 --optimize_or_like_chain 1 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 25 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 36192420 --use_uncompressed_cache 1 --min_bytes_to_use_direct_io 6541293247 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method pread_threadpool --remote_filesystem_read_method read --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 50 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 128Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 1 --compile_sort_description 1 --merge_tree_coarse_index_granularity 21 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 5244846134 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 1189983565 --min_compress_block_size 3063502 --max_compress_block_size 1838624 --merge_tree_compact_parts_min_granules_to_multibuffer_read 42 --optimize_sorting_by_input_stream_properties 1 --http_response_buffer_size 8139777 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 3 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.65 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 0 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 1 --max_parsing_threads 1 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 1 --enable_vertical_final 1
2025-04-01 17:46:04
2025-04-01 17:46:04 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 3438478178 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 17218000 --merge_max_block_size 4659 --index_granularity 39897 --min_bytes_for_wide_part 1073741824 --marks_compress_block_size 69552 --primary_key_compress_block_size 91799 --replace_long_file_name_to_hash 0 --max_file_name_length 128 --min_bytes_for_full_part_storage 342185656 --compact_parts_max_bytes_to_buffer 263592454 --compact_parts_max_granules_to_buffer 1 --compact_parts_merge_max_bytes_to_prefetch_part 6838531 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 1 --old_parts_lifetime 222 --prewarm_mark_cache 0 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 1 --use_primary_key_cache 1 --prewarm_primary_key_cache 1
2025-04-01 17:46:04
2025-04-01 17:46:04 Database: test_89ntz7bn
2025-04-01 17:46:04 01903_correct_block_size_prediction_with_default: [ FAIL ] 73.22 sec.
2025-04-01 17:46:04 Reason: return code: 210
2025-04-01 17:46:04 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-01 17:46:04 (query: SELECT uniq(15Id) FROM test_extract PREWHERE 15Id < 4 SETTINGS max_threads=1)
2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:04
2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:04
2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:04
2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:04
2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:04
2025-04-01 17:46:04 , result:
2025-04-01 17:46:04
2025-04-01 17:46:04 8
2025-04-01 17:46:04 8
2025-04-01 17:46:04 1
2025-04-01 17:46:04
2025-04-01 17:46:04 stdout:
2025-04-01 17:46:04 8
2025-04-01 17:46:04 8
2025-04-01 17:46:04 1
2025-04-01 17:46:04
2025-04-01 17:46:04
2025-04-01 17:46:04
2025-04-01 17:46:04 Database: test_ofr5rq7t
2025-04-01 17:46:05 03036_test_parquet_bloom_filter_push_down: [ FAIL ] 17.84 sec.
2025-04-01 17:46:05 Reason: having stderror:
2025-04-01 17:46:05 Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:58606): (localhost:9000, ::1, local address: [::1]:58606). (NETWORK_ERROR)
2025-04-01 17:46:05
2025-04-01 17:46:05 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:05
2025-04-01 17:46:05 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:05
2025-04-01 17:46:05 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:05
2025-04-01 17:46:05 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:05
2025-04-01 17:46:05
2025-04-01 17:46:05 stdout:
2025-04-01 17:46:05 1000
2025-04-01 17:46:05 bloom filter is off, all row groups should be read
2025-04-01 17:46:05 expect rows_read = select count()
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "AZSR",
2025-04-01 17:46:05 "flba": "WNMM"
2025-04-01 17:46:05 },
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 2,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 1000,
2025-04-01 17:46:05 "bytes_read": 47419
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 bloom filter is on, some row groups should be skipped
2025-04-01 17:46:05 expect rows_read much less than select count()
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "AZSR",
2025-04-01 17:46:05 "flba": "WNMM"
2025-04-01 17:46:05 },
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 2,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 464,
2025-04-01 17:46:05 "bytes_read": 21703
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 bloom filter is on, but where predicate contains data from 2 row groups out of 3.
2025-04-01 17:46:05 Rows read should be less than select count, but greater than previous selects
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 },
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "ZHZK",
2025-04-01 17:46:05 "flba": "HRWD"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 2,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 536,
2025-04-01 17:46:05 "bytes_read": 25708
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 bloom filter is on, but where predicate contains data from all row groups
2025-04-01 17:46:05 expect rows_read = select count()
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 },
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "OKAI",
2025-04-01 17:46:05 "flba": "UXGT"
2025-04-01 17:46:05 },
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "ZHZK",
2025-04-01 17:46:05 "flba": "HRWD"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 3,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 1000,
2025-04-01 17:46:05 "bytes_read": 47419
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 IN check
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 },
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "ZHZK",
2025-04-01 17:46:05 "flba": "HRWD"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 2,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 536,
2025-04-01 17:46:05 "bytes_read": 25708
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 tuple in case, bf is off.
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 1000,
2025-04-01 17:46:05 "bytes_read": 47419
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 tuple in case, bf is on.
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 464,
2025-04-01 17:46:05 "bytes_read": 21703
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 complex tuple in case, bf is off
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 1000,
2025-04-01 17:46:05 "bytes_read": 47419
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 complex tuple in case, bf is on
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "string": "PFJH",
2025-04-01 17:46:05 "flba": "GKJC"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 464,
2025-04-01 17:46:05 "bytes_read": 21703
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 complex tuple in case, bf is on. Non existent
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [],
2025-04-01 17:46:05 "rows": 0,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 0,
2025-04-01 17:46:05 "bytes_read": 0
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 Bloom filter for json column. BF is off
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "json": "{\"key\":38, \"value\":\"NXONM\"}"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 1000,
2025-04-01 17:46:05 "bytes_read": 47419
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 Bloom filter for json column. BF is on
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "json": "{\"key\":38, \"value\":\"NXONM\"}"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 72,
2025-04-01 17:46:05 "bytes_read": 4005
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 Bloom filter for ipv4 column. BF is off
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "json": "{\"key\":38, \"value\":\"NXONM\"}"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 1000,
2025-04-01 17:46:05 "bytes_read": 47419
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 Bloom filter for ipv4 column. BF is on
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "json": "{\"key\":38, \"value\":\"NXONM\"}"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 72,
2025-04-01 17:46:05 "bytes_read": 4005
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 Bloom filter for ipv4 column. BF is on. Specified in the schema
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "ipv4": "0.0.1.143"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 72,
2025-04-01 17:46:05 "bytes_read": 4005
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 Bloom filter on 64 bit column read as ipv4. We explicitly deny it, should read all rg
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "uint64_logical": "22.230.220.164"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 1000,
2025-04-01 17:46:05 "bytes_read": 47419
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 BF off for parquet uint64 logical type. Should read everything
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "json": "{\"key\":683, \"value\":\"YKCPD\"}"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 1000,
2025-04-01 17:46:05 "bytes_read": 47419
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 BF on for parquet uint64 logical type. Uint64 is stored as a signed int 64, but with logical annotation. Make sure a value greater than int64 can be queried
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "data": [
2025-04-01 17:46:05 {
2025-04-01 17:46:05 "json": "{\"key\":683, \"value\":\"YKCPD\"}"
2025-04-01 17:46:05 }
2025-04-01 17:46:05 ],
2025-04-01 17:46:05 "rows": 1,
2025-04-01 17:46:05 "statistics": {
2025-04-01 17:46:05 "rows_read": 464,
2025-04-01 17:46:05 "bytes_read": 21711
2025-04-01 17:46:05 }
2025-04-01 17:46:05 }
2025-04-01 17:46:05 Uint16 is stored as physical type int32 with bidwidth = 16 and sign = false. Make sure a value greater than int16 can be queried. BF is on.
2025-04-01 17:46:05 BF off for parquet int8 logical type. Should read everything
2025-04-01 17:46:05 BF on for parquet int8 logical type. Should skip row groups
2025-04-01 17:46:05 Invalid column conversion with in operation. String type can not be hashed against parquet int64 physical type. Should read everything
2025-04-01 17:46:05 Transformations on key column shall not be allowed. Should read everything
2025-04-01 17:46:05
2025-04-01 17:46:05 Settings used in the test: --max_insert_threads 1 --group_by_two_level_threshold 176251 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 0 --fsync_metadata 0 --output_format_parallel_formatting 0 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 6090155 --max_read_buffer_size 661397 --prefer_localhost_replica 0 --max_block_size 37458 --max_joined_block_size_rows 59705 --max_threads 1 --optimize_append_index 0 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 0 --optimize_or_like_chain 1 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 36 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 777581 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 8378955596 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method read --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 2 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 128Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 8Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 100Mi --compile_aggregate_expressions 1 --compile_sort_description 1 --merge_tree_coarse_index_granularity 19 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 433165382 --min_compress_block_size 1014462 --max_compress_block_size 1822062 --merge_tree_compact_parts_min_granules_to_multibuffer_read 39 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 2615188 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.67 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 1 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 0 --max_parsing_threads 1 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table false --output_format_native_write_json_as_string 1 --enable_vertical_final 0
2025-04-01 17:46:05
2025-04-01 17:46:05 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 10737418240 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 100 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 670404 --merge_max_block_size 7911 --index_granularity 39340 --min_bytes_for_wide_part 0 --marks_compress_block_size 81766 --primary_key_compress_block_size 8481 --replace_long_file_name_to_hash 1 --max_file_name_length 119 --min_bytes_for_full_part_storage 536870912 --compact_parts_max_bytes_to_buffer 261665137 --compact_parts_max_granules_to_buffer 204 --compact_parts_merge_max_bytes_to_prefetch_part 14650012 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 41 --old_parts_lifetime 213 --prewarm_mark_cache 1 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 0 --use_primary_key_cache 1 --prewarm_primary_key_cache 0
2025-04-01 17:46:05
2025-04-01 17:46:05 Database: test_jypz7w9c
2025-04-01 17:46:07 02317_distinct_in_order_optimization_explain: [ FAIL ] 29.27 sec.
2025-04-01 17:46:07 Reason: return code: 210
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:58594): (localhost:9000, ::1, local address: [::1]:58594). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-01 17:46:07
2025-04-01 17:46:07 , result:
2025-04-01 17:46:07
2025-04-01 17:46:07 -- disable optimize_distinct_in_order
2025-04-01 17:46:07 -- distinct all primary key columns -> ordinary distinct
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- enable optimize_distinct_in_order
2025-04-01 17:46:07 -- distinct with all primary key columns -> pre-distinct optimization only
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix -> pre-distinct optimization only
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by column in distinct -> pre-distinct and final distinct optimization
2025-04-01 17:46:07 DistinctSortedTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by the same columns -> pre-distinct and final distinct optimization
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by columns are prefix of distinct columns -> pre-distinct and final distinct optimization
2025-04-01 17:46:07 DistinctSortedTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by column in distinct but non-primary key prefix -> pre-distinct and final distinct optimization
2025-04-01 17:46:07 DistinctSortedTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by column _not_ in distinct -> pre-distinct optimization only
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with non-primary key prefix -> ordinary distinct
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by column in distinct -> final distinct optimization only
2025-04-01 17:46:07 DistinctSortedTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by column _not_ in distinct -> ordinary distinct
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by _const_ column in distinct -> ordinary distinct
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- Check reading in order for distinct
2025-04-01 17:46:07 -- disabled, distinct columns match sorting key
2025-04-01 17:46:07 algorithm: Thread
2025-04-01 17:46:07 -- enabled, distinct columns match sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 -- enabled, distinct columns form prefix of sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 -- enabled, distinct columns DON't form prefix of sorting key
2025-04-01 17:46:07 algorithm: Thread
2025-04-01 17:46:07 -- enabled, distinct columns contains constant columns, non-const columns form prefix of sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 -- enabled, distinct columns contains constant columns, non-const columns match prefix of sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 -- enabled, only part of distinct columns form prefix of sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 === disable new analyzer ===
2025-04-01 17:46:07 -- enabled, check that sorting properties are propagated from ReadFromMergeTree till preliminary distinct
2025-04-01 17:46:07 Sorting: a ASC, b ASC
2025-04-01 17:46:07 Sorting: a ASC, b ASC
2025-04-01 17:46:07 -- check that reading in order optimization for ORDER BY and DISTINCT applied correctly in the same query
2025-04-01 17:46:07 -- disabled, check that sorting description for ReadFromMergeTree match ORDER BY columns
2025-04-01 17:46:07 Sorting: a ASC
2025-04-01 17:46:07 Sorting: a ASC
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization i.e. it contains columns from DISTINCT clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization, but direction used from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (1), - it contains columns from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (2), - direction used from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that disabling other 'read in order' optimizations do not disable distinct in order optimization
2025-04-01 17:46:07 === enable new analyzer ===
2025-04-01 17:46:07 -- enabled, check that sorting properties are propagated from ReadFromMergeTree till preliminary distinct
2025-04-01 17:46:07 -- disabled, check that sorting description for ReadFromMergeTree match ORDER BY columns
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization i.e. it contains columns from DISTINCT clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization, but direction used from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (1), - it contains columns from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (2), - direction used from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that disabling other 'read in order' optimizations do not disable distinct in order optimization
2025-04-01 17:46:07
2025-04-01 17:46:07 stdout:
2025-04-01 17:46:07 -- disable optimize_distinct_in_order
2025-04-01 17:46:07 -- distinct all primary key columns -> ordinary distinct
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- enable optimize_distinct_in_order
2025-04-01 17:46:07 -- distinct with all primary key columns -> pre-distinct optimization only
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix -> pre-distinct optimization only
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by column in distinct -> pre-distinct and final distinct optimization
2025-04-01 17:46:07 DistinctSortedTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by the same columns -> pre-distinct and final distinct optimization
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by columns are prefix of distinct columns -> pre-distinct and final distinct optimization
2025-04-01 17:46:07 DistinctSortedTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by column in distinct but non-primary key prefix -> pre-distinct and final distinct optimization
2025-04-01 17:46:07 DistinctSortedTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with primary key prefix and order by column _not_ in distinct -> pre-distinct optimization only
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctSortedStreamTransform
2025-04-01 17:46:07 -- distinct with non-primary key prefix -> ordinary distinct
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by column in distinct -> final distinct optimization only
2025-04-01 17:46:07 DistinctSortedTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by column _not_ in distinct -> ordinary distinct
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by _const_ column in distinct -> ordinary distinct
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 DistinctTransform
2025-04-01 17:46:07 -- Check reading in order for distinct
2025-04-01 17:46:07 -- disabled, distinct columns match sorting key
2025-04-01 17:46:07 algorithm: Thread
2025-04-01 17:46:07 -- enabled, distinct columns match sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 -- enabled, distinct columns form prefix of sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 -- enabled, distinct columns DON't form prefix of sorting key
2025-04-01 17:46:07 algorithm: Thread
2025-04-01 17:46:07 -- enabled, distinct columns contains constant columns, non-const columns form prefix of sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 -- enabled, distinct columns contains constant columns, non-const columns match prefix of sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 -- enabled, only part of distinct columns form prefix of sorting key
2025-04-01 17:46:07 algorithm: InOrder
2025-04-01 17:46:07 === disable new analyzer ===
2025-04-01 17:46:07 -- enabled, check that sorting properties are propagated from ReadFromMergeTree till preliminary distinct
2025-04-01 17:46:07 Sorting: a ASC, b ASC
2025-04-01 17:46:07 Sorting: a ASC, b ASC
2025-04-01 17:46:07 -- check that reading in order optimization for ORDER BY and DISTINCT applied correctly in the same query
2025-04-01 17:46:07 -- disabled, check that sorting description for ReadFromMergeTree match ORDER BY columns
2025-04-01 17:46:07 Sorting: a ASC
2025-04-01 17:46:07 Sorting: a ASC
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization i.e. it contains columns from DISTINCT clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization, but direction used from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (1), - it contains columns from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (2), - direction used from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that disabling other 'read in order' optimizations do not disable distinct in order optimization
2025-04-01 17:46:07 === enable new analyzer ===
2025-04-01 17:46:07 -- enabled, check that sorting properties are propagated from ReadFromMergeTree till preliminary distinct
2025-04-01 17:46:07 -- disabled, check that sorting description for ReadFromMergeTree match ORDER BY columns
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization i.e. it contains columns from DISTINCT clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization, but direction used from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (1), - it contains columns from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (2), - direction used from ORDER BY clause
2025-04-01 17:46:07 -- enabled, check that disabling other 'read in order' optimizations do not disable distinct in order optimization
2025-04-01 17:46:07
2025-04-01 17:46:07
2025-04-01 17:46:07 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 786045 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 0 --fsync_metadata 0 --output_format_parallel_formatting 0 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 4698335 --max_read_buffer_size 789110 --prefer_localhost_replica 0 --max_block_size 86889 --max_joined_block_size_rows 67122 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 0 --optimize_or_like_chain 0 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 0 --read_in_order_two_level_merge_threshold 56 --optimize_aggregation_in_order 0 --aggregation_in_order_max_block_bytes 29243934 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 9709034220 --min_bytes_to_use_mmap_io 3483143077 --local_filesystem_read_method pread --remote_filesystem_read_method read --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 2 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 32Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 0 --merge_tree_coarse_index_granularity 22 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 10711791914 --max_bytes_before_remerge_sort 1724396297 --min_compress_block_size 509181 --max_compress_block_size 961470 --merge_tree_compact_parts_min_granules_to_multibuffer_read 113 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 10285867 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction True --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.89 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 1 --cross_join_min_bytes_to_compress 1 --min_external_table_block_size_bytes 100000000 --max_parsing_threads 0 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 1 --enable_vertical_final 1
2025-04-01 17:46:07
2025-04-01 17:46:07 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 0.7563738772706338 --prefer_fetch_merged_part_size_threshold 2012106554 --vertical_merge_algorithm_min_rows_to_activate 1 --vertical_merge_algorithm_min_columns_to_activate 82 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 27699374 --merge_max_block_size 19581 --index_granularity 48176 --min_bytes_for_wide_part 0 --marks_compress_block_size 17345 --primary_key_compress_block_size 29004 --replace_long_file_name_to_hash 1 --max_file_name_length 0 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 126996618 --compact_parts_max_granules_to_buffer 256 --compact_parts_merge_max_bytes_to_prefetch_part 22889841 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 100 --old_parts_lifetime 480 --prewarm_mark_cache 1 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 0 --use_primary_key_cache 1 --prewarm_primary_key_cache 0
2025-04-01 17:46:07
2025-04-01 17:46:07 Database: test_l0nrtmbt
2025-04-01 17:46:13 02122_parallel_formatting_RowBinaryWithNames: [ FAIL ] 0.00 sec.
2025-04-01 17:46:13 Reason: server died
2025-04-01 17:46:13 Server does not respond to health check
2025-04-01 17:46:13
2025-04-01 17:46:13 01440_big_int_exotic_casts: [ FAIL ] 0.00 sec.
2025-04-01 17:46:13 Reason: server died
2025-04-01 17:46:13 Server does not respond to health check
2025-04-01 17:46:13
2025-04-01 17:46:13 Process Process-6:
2025-04-01 17:46:13 Traceback (most recent call last):
2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-01 17:46:13 self.run()
2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-01 17:46:13 self._target(*self._args, **self._kwargs)
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-01 17:46:13 return run_tests_array(*args, **kwargs)
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-01 17:46:13 stop_tests()
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:46:13 cleanup_child_processes(os.getpid())
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:46:13 child_pgid = os.getpgid(child)
2025-04-01 17:46:13 ProcessLookupError: [Errno 3] No such process
2025-04-01 17:46:13 Child processes of 1388:
2025-04-01 17:46:13 41795 /bin/sh -c pgrep --parent 1388 -a
2025-04-01 17:46:13 Process Process-4:
2025-04-01 17:46:13 Traceback (most recent call last):
2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-01 17:46:13 self.run()
2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-01 17:46:13 self._target(*self._args, **self._kwargs)
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-01 17:46:13 return run_tests_array(*args, **kwargs)
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-01 17:46:13 stop_tests()
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:46:13 cleanup_child_processes(os.getpid())
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:46:13 child_pgid = os.getpgid(child)
2025-04-01 17:46:13 ProcessLookupError: [Errno 3] No such process
2025-04-01 17:46:13 Child processes of 1384:
2025-04-01 17:46:13 41797 /bin/sh -c pgrep --parent 1384 -a
2025-04-01 17:46:13 02221_parallel_replicas_bug: [ FAIL ] 0.00 sec.
2025-04-01 17:46:13 Reason: server died
2025-04-01 17:46:13 Server does not respond to health check
2025-04-01 17:46:13
2025-04-01 17:46:13 Database: test_sczjo540
2025-04-01 17:46:13 Process Process-5:
2025-04-01 17:46:13 Traceback (most recent call last):
2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-01 17:46:13 self.run()
2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-01 17:46:13 self._target(*self._args, **self._kwargs)
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-01 17:46:13 return run_tests_array(*args, **kwargs)
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-01 17:46:13 stop_tests()
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:46:13 cleanup_child_processes(os.getpid())
2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:46:13 child_pgid = os.getpgid(child)
2025-04-01 17:46:13 ProcessLookupError: [Errno 3] No such process
2025-04-01 17:46:13 Child processes of 1385:
2025-04-01 17:46:13 41807 /bin/sh -c pgrep --parent 1385 -a
2025-04-01 17:46:14 01901_in_literal_shard_prune: [ FAIL ] 0.00 sec.
2025-04-01 17:46:14 Reason: server died
2025-04-01 17:46:14 Server does not respond to health check
2025-04-01 17:46:14
2025-04-01 17:46:14 Process Process-8:
2025-04-01 17:46:14 Traceback (most recent call last):
2025-04-01 17:46:14 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-01 17:46:14 self.run()
2025-04-01 17:46:14 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-01 17:46:14 self._target(*self._args, **self._kwargs)
2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-01 17:46:14 return run_tests_array(*args, **kwargs)
2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-01 17:46:14 stop_tests()
2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:46:14 cleanup_child_processes(os.getpid())
2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:46:14 child_pgid = os.getpgid(child)
2025-04-01 17:46:14 ProcessLookupError: [Errno 3] No such process
2025-04-01 17:46:14 Child processes of 1391:
2025-04-01 17:46:14 41813 /bin/sh -c pgrep --parent 1391 -a
2025-04-01 17:46:14 02565_update_empty_nested: [ FAIL ] 0.00 sec.
2025-04-01 17:46:14 Reason: server died
2025-04-01 17:46:14 Server does not respond to health check
2025-04-01 17:46:14
2025-04-01 17:46:14 Process Process-7:
2025-04-01 17:46:14 Traceback (most recent call last):
2025-04-01 17:46:14 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-01 17:46:14 self.run()
2025-04-01 17:46:14 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-01 17:46:14 self._target(*self._args, **self._kwargs)
2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-01 17:46:14 return run_tests_array(*args, **kwargs)
2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-01 17:46:14 stop_tests()
2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:46:14 cleanup_child_processes(os.getpid())
2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:46:14 child_pgid = os.getpgid(child)
2025-04-01 17:46:14 ProcessLookupError: [Errno 3] No such process
2025-04-01 17:46:14 Child processes of 1390:
2025-04-01 17:46:14 41819 /bin/sh -c pgrep --parent 1390 -a
2025-04-01 17:46:15 01115_prewhere_array_join: [ FAIL ] 0.00 sec.
2025-04-01 17:46:15 Reason: server died
2025-04-01 17:46:15 Server does not respond to health check
2025-04-01 17:46:15
2025-04-01 17:46:15 Process Process-3:
2025-04-01 17:46:15 Traceback (most recent call last):
2025-04-01 17:46:15 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-01 17:46:15 self.run()
2025-04-01 17:46:15 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-01 17:46:15 self._target(*self._args, **self._kwargs)
2025-04-01 17:46:15 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-01 17:46:15 return run_tests_array(*args, **kwargs)
2025-04-01 17:46:15 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-01 17:46:15 stop_tests()
2025-04-01 17:46:15 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:46:15 cleanup_child_processes(os.getpid())
2025-04-01 17:46:15 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:46:15 child_pgid = os.getpgid(child)
2025-04-01 17:46:15 ProcessLookupError: [Errno 3] No such process
2025-04-01 17:46:15 Child processes of 1383:
2025-04-01 17:46:15 41825 /bin/sh -c pgrep --parent 1383 -a
2025-04-01 17:46:17 01047_no_alias_columns_with_table_aliases: [ FAIL ] 0.00 sec.
2025-04-01 17:46:17 Reason: server died
2025-04-01 17:46:17 Server does not respond to health check
2025-04-01 17:46:17
2025-04-01 17:46:17 Process Process-10:
2025-04-01 17:46:17 Traceback (most recent call last):
2025-04-01 17:46:17 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-01 17:46:17 self.run()
2025-04-01 17:46:17 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-01 17:46:17 self._target(*self._args, **self._kwargs)
2025-04-01 17:46:17 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-01 17:46:17 return run_tests_array(*args, **kwargs)
2025-04-01 17:46:17 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-01 17:46:17 stop_tests()
2025-04-01 17:46:17 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:46:17 cleanup_child_processes(os.getpid())
2025-04-01 17:46:17 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:46:17 child_pgid = os.getpgid(child)
2025-04-01 17:46:17 ProcessLookupError: [Errno 3] No such process
2025-04-01 17:46:17 Child processes of 1397:
2025-04-01 17:46:17 41831 /bin/sh -c pgrep --parent 1397 -a
2025-04-01 17:59:04 Hung check failed: [Errno 111] Connection refused
2025-04-01 17:59:04 Server died, terminating all processes...
2025-04-01 17:59:04 Running 130 stateless tests (MainProcess).
2025-04-01 17:59:04 Child processes of 1367:
2025-04-01 17:59:04 1370 python3 /usr/bin/clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 2 --run-by-hash-total 4 --report-logs-stats
2025-04-01 17:59:04 1376 python3 /usr/bin/clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 2 --run-by-hash-total 4 --report-logs-stats
2025-04-01 17:59:04 41839 /bin/sh -c pgrep --parent 1367 -a
2025-04-01 17:59:04 Traceback (most recent call last):
2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 3716, in
2025-04-01 17:59:04 main(args)
2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 3063, in main
2025-04-01 17:59:04 total_tests_run += do_run_tests(args.jobs, test_suite)
2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 2697, in do_run_tests
2025-04-01 17:59:04 run_tests_array(
2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 2343, in run_tests_array
2025-04-01 17:59:04 stop_tests()
2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-01 17:59:04 cleanup_child_processes(os.getpid())
2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-01 17:59:04 child_pgid = os.getpgid(child)
2025-04-01 17:59:04 ProcessLookupError: [Errno 3] No such process
+ set -e
+ echo 'Files in current directory'
+ ls -la ./
Files in current directory
total 127172
drwxr-xr-x 1 root root 4096 Apr 1 17:45 .
drwxr-xr-x 1 root root 4096 Apr 1 17:45 ..
lrwxrwxrwx 1 root root 7 Sep 11 2024 bin -> usr/bin
drwxr-xr-x 2 root root 4096 Apr 18 2022 boot
drwxr-xr-x 14 root root 3840 Apr 1 17:39 dev
-rwxr-xr-x 1 root root 0 Apr 1 17:39 .dockerenv
drwxr-xr-x 1 root root 4096 Apr 1 17:39 etc
drwxr-xr-x 2 root root 4096 Apr 18 2022 home
lrwxrwxrwx 1 root root 7 Sep 11 2024 lib -> usr/lib
lrwxrwxrwx 1 root root 9 Sep 11 2024 lib32 -> usr/lib32
lrwxrwxrwx 1 root root 9 Sep 11 2024 lib64 -> usr/lib64
lrwxrwxrwx 1 root root 10 Sep 11 2024 libx32 -> usr/libx32
-rwxr-xr-x 1 root root 26927256 Jan 15 17:53 mc
drwxr-xr-x 2 root root 4096 Sep 11 2024 media
-rwxr-xr-x 1 root root 103174296 Jan 15 17:53 minio
drwxr-xr-x 4 root root 4096 Apr 1 17:39 minio_data
drwxr-xr-x 2 root root 4096 Sep 11 2024 mnt
drwxr-xr-x 1 root root 4096 Jan 14 17:08 opt
-rw-r--r-- 1 root root 0 Feb 14 2024 .package-cache-mutate
drwxrwxr-x 2 1000 1000 4096 Apr 1 17:39 package_folder
drwxr-x--- 2 root root 4096 Apr 1 17:41 preprocessed_configs
dr-xr-xr-x 305 root root 0 Apr 1 17:39 proc
-rw-r--r-- 1 root root 29 Apr 1 17:40 queries_02352
drwxrwxr-x 17 1000 1000 4096 Apr 1 17:36 repo
-rw-rw-r-- 1 root root 863 Jan 15 17:27 requirements.txt
drwx------ 1 root root 4096 Apr 1 17:45 root
drwxr-xr-x 1 root root 4096 Apr 1 17:39 run
lrwxrwxrwx 1 root root 8 Sep 11 2024 sbin -> usr/sbin
-rw-r--r-- 1 root root 747 Apr 1 17:39 script.gdb
-rwxrwxr-x 1 root root 10851 Jan 14 17:07 setup_export_logs.sh
drwxr-xr-x 2 root root 4096 Sep 11 2024 srv
dr-xr-xr-x 13 root root 0 Apr 1 17:39 sys
drwxrwxr-x 2 1000 1000 4096 Apr 1 17:39 test_output
drwxrwxrwt 1 root root 4096 Apr 1 17:59 tmp
drwxr-xr-x 1 root root 4096 Sep 11 2024 usr
drwxr-xr-x 1 root root 4096 Sep 11 2024 var
+ echo 'Files in root directory'
+ ls -la /
Files in root directory
total 127172
drwxr-xr-x 1 root root 4096 Apr 1 17:45 .
drwxr-xr-x 1 root root 4096 Apr 1 17:45 ..
lrwxrwxrwx 1 root root 7 Sep 11 2024 bin -> usr/bin
drwxr-xr-x 2 root root 4096 Apr 18 2022 boot
drwxr-xr-x 14 root root 3840 Apr 1 17:39 dev
-rwxr-xr-x 1 root root 0 Apr 1 17:39 .dockerenv
drwxr-xr-x 1 root root 4096 Apr 1 17:39 etc
drwxr-xr-x 2 root root 4096 Apr 18 2022 home
lrwxrwxrwx 1 root root 7 Sep 11 2024 lib -> usr/lib
lrwxrwxrwx 1 root root 9 Sep 11 2024 lib32 -> usr/lib32
lrwxrwxrwx 1 root root 9 Sep 11 2024 lib64 -> usr/lib64
lrwxrwxrwx 1 root root 10 Sep 11 2024 libx32 -> usr/libx32
-rwxr-xr-x 1 root root 26927256 Jan 15 17:53 mc
drwxr-xr-x 2 root root 4096 Sep 11 2024 media
-rwxr-xr-x 1 root root 103174296 Jan 15 17:53 minio
drwxr-xr-x 4 root root 4096 Apr 1 17:39 minio_data
drwxr-xr-x 2 root root 4096 Sep 11 2024 mnt
drwxr-xr-x 1 root root 4096 Jan 14 17:08 opt
-rw-r--r-- 1 root root 0 Feb 14 2024 .package-cache-mutate
drwxrwxr-x 2 1000 1000 4096 Apr 1 17:39 package_folder
drwxr-x--- 2 root root 4096 Apr 1 17:41 preprocessed_configs
dr-xr-xr-x 305 root root 0 Apr 1 17:39 proc
-rw-r--r-- 1 root root 29 Apr 1 17:40 queries_02352
drwxrwxr-x 17 1000 1000 4096 Apr 1 17:36 repo
-rw-rw-r-- 1 root root 863 Jan 15 17:27 requirements.txt
drwx------ 1 root root 4096 Apr 1 17:45 root
drwxr-xr-x 1 root root 4096 Apr 1 17:39 run
lrwxrwxrwx 1 root root 8 Sep 11 2024 sbin -> usr/sbin
-rw-r--r-- 1 root root 747 Apr 1 17:39 script.gdb
-rwxrwxr-x 1 root root 10851 Jan 14 17:07 setup_export_logs.sh
drwxr-xr-x 2 root root 4096 Sep 11 2024 srv
dr-xr-xr-x 13 root root 0 Apr 1 17:39 sys
drwxrwxr-x 2 1000 1000 4096 Apr 1 17:39 test_output
drwxrwxrwt 1 root root 4096 Apr 1 17:59 tmp
drwxr-xr-x 1 root root 4096 Sep 11 2024 usr
drwxr-xr-x 1 root root 4096 Sep 11 2024 var
+ clickhouse-client -q 'system flush logs'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ :
+ stop_logs_replication
+ echo 'Detach all logs replication'
Detach all logs replication
+ clickhouse-client --query 'select database||'\''.'\''||table from system.tables where database = '\''system'\'' and (table like '\''%_sender'\'' or table like '\''%_watcher'\'')'
+ tee /dev/stderr
+ timeout --preserve-status --signal TERM --kill-after 5m 15m xargs -n1 -r -i clickhouse-client --query 'drop table {}'
xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ logs_saver_client_options='--max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0'
+ failed_to_save_logs=0
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.query_log into outfile '\''/test_output/query_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.zookeeper_log into outfile '\''/test_output/zookeeper_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.trace_log into outfile '\''/test_output/trace_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.transactions_info_log into outfile '\''/test_output/transactions_info_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.metric_log into outfile '\''/test_output/metric_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.blob_storage_log into outfile '\''/test_output/blob_storage_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.error_log into outfile '\''/test_output/error_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.query_metric_log into outfile '\''/test_output/query_metric_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ sleep 1
+ clickhouse-client -q 'SYSTEM FLUSH ASYNC INSERT QUEUE'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ :
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'SELECT log FROM minio_audit_logs ORDER BY log.time INTO OUTFILE '\''/test_output/minio_audit_logs.jsonl.zst'\'' FORMAT JSONEachRow'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ :
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'SELECT log FROM minio_server_logs ORDER BY log.time INTO OUTFILE '\''/test_output/minio_server_logs.jsonl.zst'\'' FORMAT JSONEachRow'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ :
+ sudo clickhouse stop
==41896==Installed the sigaction for signal 11
==41896==Installed the sigaction for signal 7
==41896==Installed the sigaction for signal 8
==41896==Using llvm-symbolizer found at: /usr/bin/llvm-symbolizer-19
***** Running under ThreadSanitizer v3 (pid 41896) *****
ThreadSanitizer: growing sync allocator: 0 out of 1048576*1024
ThreadSanitizer: growing heap block allocator: 0 out of 262144*4096
ThreadSanitizer: growing heap block allocator: 1 out of 262144*4096
ThreadSanitizer: growing sync allocator: 1 out of 1048576*1024
ThreadSanitizer: growing heap block allocator: 2 out of 262144*4096
ThreadSanitizer: growing sync allocator: 2 out of 1048576*1024
ThreadSanitizer: growing sync allocator: 3 out of 1048576*1024
ThreadSanitizer: growing sync allocator: 4 out of 1048576*1024
ThreadSanitizer: growing heap block allocator: 3 out of 262144*4096
ThreadSanitizer: growing sync allocator: 5 out of 1048576*1024
ThreadSanitizer: growing sync allocator: 6 out of 1048576*1024
/var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 428.
The process with pid = 428 does not exist.
Stats: SizeClassAllocator64: 10M mapped (1M rss) in 20235 allocations; remains 5797
01 ( 16): mapped: 256K allocs: 1024 frees: 768 inuse: 256 num_freed_chunks 16128 avail: 16384 rss: 16K releases: 0 last released: 0K region: 0x720400000000
02 ( 32): mapped: 256K allocs: 7168 frees: 6656 inuse: 512 num_freed_chunks 7680 avail: 8192 rss: 176K releases: 0 last released: 0K region: 0x720800000000
03 ( 48): mapped: 256K allocs: 5632 frees: 4608 inuse: 1024 num_freed_chunks 4437 avail: 5461 rss: 252K releases: 0 last released: 0K region: 0x720c00000000
04 ( 64): mapped: 256K allocs: 1792 frees: 1408 inuse: 384 num_freed_chunks 3712 avail: 4096 rss: 108K releases: 0 last released: 0K region: 0x721000000000
05 ( 80): mapped: 256K allocs: 384 frees: 128 inuse: 256 num_freed_chunks 3020 avail: 3276 rss: 28K releases: 0 last released: 0K region: 0x721400000000
06 ( 96): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 2474 avail: 2730 rss: 20K releases: 0 last released: 0K region: 0x721800000000
07 ( 112): mapped: 256K allocs: 384 frees: 128 inuse: 256 num_freed_chunks 2084 avail: 2340 rss: 36K releases: 0 last released: 0K region: 0x721c00000000
08 ( 128): mapped: 256K allocs: 640 frees: 384 inuse: 256 num_freed_chunks 1792 avail: 2048 rss: 80K releases: 0 last released: 0K region: 0x722000000000
09 ( 144): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1692 avail: 1820 rss: 8K releases: 0 last released: 0K region: 0x722400000000
10 ( 160): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1510 avail: 1638 rss: 8K releases: 0 last released: 0K region: 0x722800000000
11 ( 176): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1361 avail: 1489 rss: 4K releases: 0 last released: 0K region: 0x722c00000000
12 ( 192): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1237 avail: 1365 rss: 16K releases: 0 last released: 0K region: 0x723000000000
13 ( 208): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1132 avail: 1260 rss: 4K releases: 0 last released: 0K region: 0x723400000000
14 ( 224): mapped: 256K allocs: 512 frees: 256 inuse: 256 num_freed_chunks 914 avail: 1170 rss: 92K releases: 0 last released: 0K region: 0x723800000000
15 ( 240): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 836 avail: 1092 rss: 40K releases: 0 last released: 0K region: 0x723c00000000
16 ( 256): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 768 avail: 1024 rss: 40K releases: 0 last released: 0K region: 0x724000000000
17 ( 320): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 691 avail: 819 rss: 16K releases: 0 last released: 0K region: 0x724400000000
18 ( 384): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 554 avail: 682 rss: 8K releases: 0 last released: 0K region: 0x724800000000
19 ( 448): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 457 avail: 585 rss: 8K releases: 0 last released: 0K region: 0x724c00000000
20 ( 512): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 384 avail: 512 rss: 4K releases: 0 last released: 0K region: 0x725000000000
21 ( 640): mapped: 256K allocs: 102 frees: 0 inuse: 102 num_freed_chunks 307 avail: 409 rss: 4K releases: 0 last released: 0K region: 0x725400000000
22 ( 768): mapped: 256K allocs: 85 frees: 0 inuse: 85 num_freed_chunks 256 avail: 341 rss: 4K releases: 0 last released: 0K region: 0x725800000000
23 ( 896): mapped: 256K allocs: 73 frees: 0 inuse: 73 num_freed_chunks 219 avail: 292 rss: 8K releases: 0 last released: 0K region: 0x725c00000000
24 ( 1024): mapped: 256K allocs: 64 frees: 0 inuse: 64 num_freed_chunks 192 avail: 256 rss: 8K releases: 0 last released: 0K region: 0x726000000000
25 ( 1280): mapped: 256K allocs: 204 frees: 102 inuse: 102 num_freed_chunks 102 avail: 204 rss: 252K releases: 0 last released: 0K region: 0x726400000000
26 ( 1536): mapped: 256K allocs: 42 frees: 0 inuse: 42 num_freed_chunks 128 avail: 170 rss: 4K releases: 0 last released: 0K region: 0x726800000000
27 ( 1792): mapped: 256K allocs: 36 frees: 0 inuse: 36 num_freed_chunks 110 avail: 146 rss: 4K releases: 0 last released: 0K region: 0x726c00000000
28 ( 2048): mapped: 256K allocs: 32 frees: 0 inuse: 32 num_freed_chunks 96 avail: 128 rss: 8K releases: 0 last released: 0K region: 0x727000000000
29 ( 2560): mapped: 256K allocs: 25 frees: 0 inuse: 25 num_freed_chunks 77 avail: 102 rss: 20K releases: 0 last released: 0K region: 0x727400000000
30 ( 3072): mapped: 256K allocs: 21 frees: 0 inuse: 21 num_freed_chunks 64 avail: 85 rss: 8K releases: 0 last released: 0K region: 0x727800000000
31 ( 3584): mapped: 256K allocs: 18 frees: 0 inuse: 18 num_freed_chunks 55 avail: 73 rss: 12K releases: 0 last released: 0K region: 0x727c00000000
32 ( 4096): mapped: 256K allocs: 16 frees: 0 inuse: 16 num_freed_chunks 48 avail: 64 rss: 12K releases: 0 last released: 0K region: 0x728000000000
33 ( 5120): mapped: 256K allocs: 12 frees: 0 inuse: 12 num_freed_chunks 39 avail: 51 rss: 16K releases: 0 last released: 0K region: 0x728400000000
34 ( 6144): mapped: 256K allocs: 10 frees: 0 inuse: 10 num_freed_chunks 32 avail: 42 rss: 12K releases: 0 last released: 0K region: 0x728800000000
35 ( 7168): mapped: 256K allocs: 9 frees: 0 inuse: 9 num_freed_chunks 27 avail: 36 rss: 52K releases: 0 last released: 0K region: 0x728c00000000
36 ( 8192): mapped: 256K allocs: 8 frees: 0 inuse: 8 num_freed_chunks 24 avail: 32 rss: 16K releases: 0 last released: 0K region: 0x729000000000
37 ( 10240): mapped: 256K allocs: 6 frees: 0 inuse: 6 num_freed_chunks 19 avail: 25 rss: 12K releases: 0 last released: 0K region: 0x729400000000
38 ( 12288): mapped: 256K allocs: 5 frees: 0 inuse: 5 num_freed_chunks 16 avail: 21 rss: 12K releases: 0 last released: 0K region: 0x729800000000
39 ( 14336): mapped: 256K allocs: 4 frees: 0 inuse: 4 num_freed_chunks 14 avail: 18 rss: 16K releases: 0 last released: 0K region: 0x729c00000000
42 ( 24576): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 8 avail: 10 rss: 16K releases: 0 last released: 0K region: 0x72a800000000
43 ( 28672): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 7 avail: 9 rss: 28K releases: 0 last released: 0K region: 0x72ac00000000
44 ( 32768): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 6 avail: 8 rss: 32K releases: 0 last released: 0K region: 0x72b000000000
45 ( 40960): mapped: 256K allocs: 1 frees: 0 inuse: 1 num_freed_chunks 5 avail: 6 rss: 4K releases: 0 last released: 0K region: 0x72b400000000
Stats: LargeMmapAllocator: allocated 1 times, remains 0 (0 K) max 1 M; by size logs: 20:1;
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ kill 1344
+ rg -Fa '' /var/log/clickhouse-server/clickhouse-server.log
2025.04.01 17:45:29.274291 [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Logical error: 'Replica info is not initialized'.
2025.04.01 17:45:29.375723 [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Stack trace (when copying this message, always include the lines below):
2025.04.01 17:45:29.377757 [ 40158 ] {} BaseDaemon: ########## Short fault info ############
2025.04.01 17:45:29.378085 [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2230) Received signal 6
2025.04.01 17:45:29.378470 [ 40158 ] {} BaseDaemon: Signal description: Aborted
2025.04.01 17:45:29.378820 [ 40158 ] {} BaseDaemon:
2025.04.01 17:45:29.379100 [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x000055dfaf3e1a3c 0x000055dfaf3ded6e 0x000055dfaf3e0c0d 0x000055dfb3a56bfe 0x000055dfb3536f8b 0x000055dfb3a5684d 0x000055dfb355ba47 0x000055dfb354b54c 0x000055dfb354a8cc 0x000055dfb354a5aa 0x000055dfb354936a 0x000055dfa75187ee 0x000055dfa75209dc 0x000055df9eb31428 0x00007f1bbdd2cac3 0x00007f1bbddbe850
2025.04.01 17:45:29.379670 [ 40158 ] {} BaseDaemon: ########################################
2025.04.01 17:45:29.380298 [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2230) (query_id: 549c2d11-5172-44cf-9edd-d2571bb9f6a9) (query: INSERT INTO distributed_01099_b SELECT * FROM urlCluster('test_cluster_two_shards', 'http://localhost:8123/?query=select+{1,2,3}+format+TSV', 'TSV', 's String');) Received signal Aborted (6)
2025.04.01 17:45:29.381012 [ 40158 ] {} BaseDaemon:
2025.04.01 17:45:29.381547 [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x000055dfaf3e1a3c 0x000055dfaf3ded6e 0x000055dfaf3e0c0d 0x000055dfb3a56bfe 0x000055dfb3536f8b 0x000055dfb3a5684d 0x000055dfb355ba47 0x000055dfb354b54c 0x000055dfb354a8cc 0x000055dfb354a5aa 0x000055dfb354936a 0x000055dfa75187ee 0x000055dfa75209dc 0x000055df9eb31428 0x00007f1bbdd2cac3 0x00007f1bbddbe850
2025.04.01 17:45:29.582232 [ 40158 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()
2025.04.01 17:45:29.582934 [ 40158 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d
2025.04.01 17:45:29.745561 [ 40158 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422
2025.04.01 17:45:32.123765 [ 40158 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6
2025.04.01 17:45:32.261573 [ 40158 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb
2025.04.01 17:45:32.261814 [ 40158 ] {} BaseDaemon: 4. ? @ 0x00007f1bbdcda520
2025.04.01 17:45:32.262082 [ 40158 ] {} BaseDaemon: 5. ? @ 0x00007f1bbdd2e9fd
2025.04.01 17:45:32.262239 [ 40158 ] {} BaseDaemon: 6. ? @ 0x00007f1bbdcda476
2025.04.01 17:45:32.262493 [ 40158 ] {} BaseDaemon: 7. ? @ 0x00007f1bbdcc07f3
2025.04.01 17:45:32.409562 [ 40158 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267
2025.04.01 17:45:32.688103 [ 40158 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a
2025.04.01 17:45:32.974951 [ 40158 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)
2025.04.01 17:45:32.975240 [ 40158 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff
2025.04.01 17:45:34.785308 [ 40158 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade
2025.04.01 17:45:34.920448 [ 40158 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba
2025.04.01 17:45:35.364440 [ 40158 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c
2025.04.01 17:45:35.710991 [ 40158 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e
2025.04.01 17:45:36.167573 [ 40158 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d
2025.04.01 17:45:36.331098 [ 40158 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe
2025.04.01 17:45:36.417293 [ 40158 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b
2025.04.01 17:45:36.515182 [ 40158 ] {} BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d
2025.04.01 17:45:36.572278 [ 40158 ] {} BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*)
2025.04.01 17:45:36.572595 [ 40158 ] {} BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47
2025.04.01 17:45:36.688121 [ 40158 ] {} BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c
2025.04.01 17:45:36.790313 [ 40158 ] {} BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long)
2025.04.01 17:45:36.790775 [ 40158 ] {} BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc
2025.04.01 17:45:36.901081 [ 40158 ] {} BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa
2025.04.01 17:45:36.972964 [ 40158 ] {} BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool)
2025.04.01 17:45:36.973222 [ 40158 ] {} BaseDaemon: 23.1. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:89: operator()
2025.04.01 17:45:36.973358 [ 40158 ] {} BaseDaemon: 23.2. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?
2025.04.01 17:45:36.973992 [ 40158 ] {} BaseDaemon: 23.3. inlined from ./contrib/llvm-project/libcxx/include/tuple:1789: _ZNSt3__118__apply_tuple_implB6v15007IRZN2DB25CompletedPipelineExecutor7executeEvE3$_0RNS_5tupleIJEEETpTnmJEEEDcOT_OT0_NS_15__tuple_indicesIJXspT1_EEEE
2025.04.01 17:45:36.974829 [ 40158 ] {} BaseDaemon: 23.4. inlined from ./contrib/llvm-project/libcxx/include/tuple:1798: decltype(auto) std::apply[abi:v15007]&>(DB::CompletedPipelineExecutor::execute()::$_0&, std::tuple<>&)
2025.04.01 17:45:36.975028 [ 40158 ] {} BaseDaemon: 23.5. inlined from ./src/Common/ThreadPool.h:311: operator()
2025.04.01 17:45:36.975334 [ 40158 ] {} BaseDaemon: 23.6. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?
2025.04.01 17:45:36.975646 [ 40158 ] {} BaseDaemon: 23.7. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:479: ?
2025.04.01 17:45:36.975876 [ 40158 ] {} BaseDaemon: 23.8. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:235: ?
2025.04.01 17:45:36.976343 [ 40158 ] {} BaseDaemon: 23. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x000000001c52036a
2025.04.01 17:45:37.070390 [ 40158 ] {} BaseDaemon: 24.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:848: ?
2025.04.01 17:45:37.070675 [ 40158 ] {} BaseDaemon: 24.1. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:1197: ?
2025.04.01 17:45:37.070894 [ 40158 ] {} BaseDaemon: 24. ./build_docker/./src/Common/ThreadPool.cpp:785: ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x00000000104ef7ee
2025.04.01 17:45:37.228767 [ 40158 ] {} BaseDaemon: 25.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:359: ?
2025.04.01 17:45:37.229273 [ 40158 ] {} BaseDaemon: 25.1. inlined from ./contrib/llvm-project/libcxx/include/thread:284: void std::__thread_execute[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*, 2ul>(std::tuple>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>&, std::__tuple_indices<2ul>)
2025.04.01 17:45:37.229646 [ 40158 ] {} BaseDaemon: 25. ./contrib/llvm-project/libcxx/include/thread:295: void* std::__thread_proxy[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>>(void*) @ 0x00000000104f79dc
2025.04.01 17:45:38.524260 [ 40158 ] {} BaseDaemon: 26. __tsan_thread_start_func @ 0x0000000007b08428
2025.04.01 17:45:38.524385 [ 40158 ] {} BaseDaemon: 27. ? @ 0x00007f1bbdd2cac3
2025.04.01 17:45:38.524515 [ 40158 ] {} BaseDaemon: 28. ? @ 0x00007f1bbddbe850
2025.04.01 17:45:41.785737 [ 40158 ] {} BaseDaemon: Integrity check of the executable successfully passed (checksum: CAD5C347B3B532C05434FCAB6F84FB12)
2025.04.01 17:45:49.494525 [ 40158 ] {} BaseDaemon: This ClickHouse version is not official and should be upgraded to the official build.
2025.04.01 17:45:49.496511 [ 40158 ] {} BaseDaemon: Changed settings: min_compress_block_size = 838722, max_compress_block_size = 1814485, max_block_size = 43866, min_external_table_block_size_bytes = 0, max_joined_block_size_rows = 36602, max_insert_threads = 3, max_threads = 1, max_parsing_threads = 1, max_read_buffer_size = 772261, connect_timeout_with_failover_ms = 2000, connect_timeout_with_failover_secure_ms = 3000, idle_connection_timeout = 36000, s3_max_get_rps = 1000000, s3_max_get_burst = 2000000, s3_max_put_rps = 1000000, s3_max_put_burst = 2000000, s3_check_objects_after_upload = true, use_uncompressed_cache = true, max_remote_read_network_bandwidth = 1000000000000, max_remote_write_network_bandwidth = 1000000000000, max_local_read_bandwidth = 1000000000000, max_local_write_bandwidth = 1000000000000, stream_like_engine_allow_direct_select = true, enable_multiple_prewhere_read_steps = false, replication_wait_for_inactive_replica_timeout = 30, compile_sort_description = false, group_by_two_level_threshold = 756728, group_by_two_level_threshold_bytes = 14099342, allow_nonconst_timezone_arguments = true, parallel_distributed_insert_select = 1, input_format_parallel_parsing = false, min_chunk_bytes_for_parallel_parsing = 9298309, merge_tree_coarse_index_granularity = 16, min_bytes_to_use_direct_io = 10737418240, min_bytes_to_use_mmap_io = 10737418240, log_queries = true, insert_quorum_timeout = 60000, merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability = 0.4000000059604645, http_response_buffer_size = 9705764, fsync_metadata = true, query_plan_join_swap_table = true, http_send_timeout = 60., http_receive_timeout = 60., use_index_for_in_with_subqueries_max_values = 1000000000, opentelemetry_start_trace_probability = 0.10000000149011612, max_rows_to_read = 20000000, max_bytes_to_read = 1000000000000, max_bytes_to_read_leaf = 1000000000000, max_rows_to_group_by = 10000000000, max_bytes_before_external_group_by = 10737418240, max_rows_to_sort = 10000000000, max_bytes_to_sort = 10000000000, prefer_external_sort_block_bytes = 100000000, max_bytes_before_external_sort = 10737418240, max_bytes_before_remerge_sort = 936605911, max_result_rows = 1000000000, max_result_bytes = 1000000000, max_execution_time = 600., max_execution_time_leaf = 600., max_execution_speed = 100000000000, max_execution_speed_bytes = 10000000000000, timeout_before_checking_execution_speed = 300., max_estimated_execution_time = 600., max_columns_to_read = 20000, max_temporary_columns = 20000, max_temporary_non_const_columns = 20000, max_rows_in_set = 10000000000, max_bytes_in_set = 10000000000, max_rows_in_join = 10000000000, max_bytes_in_join = 10000000000, cross_join_min_rows_to_compress = 1, cross_join_min_bytes_to_compress = 1, max_rows_to_transfer = 1000000000, max_bytes_to_transfer = 1000000000, max_rows_in_distinct = 10000000000, max_bytes_in_distinct = 10000000000, max_memory_usage = 5000000000, max_memory_usage_for_user = 32000000000, max_untracked_memory = 1048576, memory_profiler_step = 1048576, max_network_bandwidth = 100000000000, max_network_bytes = 1000000000000, max_network_bandwidth_for_user = 100000000000, max_network_bandwidth_for_all_users = 100000000000, max_temporary_data_on_disk_size_for_user = 100000000000, max_temporary_data_on_disk_size_for_query = 100000000000, max_backup_bandwidth = 100000000000, log_comment = '02221_parallel_replicas_bug.sh', send_logs_level = 'error', prefer_localhost_replica = false, optimize_read_in_order = false, optimize_aggregation_in_order = true, aggregation_in_order_max_block_bytes = 15394136, read_in_order_two_level_merge_threshold = 79, max_hyperscan_regexp_length = 1000000, max_hyperscan_regexp_total_length = 10000000, allow_introspection_functions = true, database_atomic_wait_for_drop_and_detach_synchronously = true, optimize_append_index = true, query_cache_max_size_in_bytes = 10000000, query_cache_max_entries = 100000, distributed_ddl_entry_format_version = 6, external_storage_max_read_rows = 10000000000, external_storage_max_read_bytes = 10000000000, local_filesystem_read_method = 'pread', merge_tree_min_bytes_per_task_for_remote_reading = 8388608, merge_tree_compact_parts_min_granules_to_multibuffer_read = 123, async_insert_busy_timeout_max_ms = 5000, enable_filesystem_cache = true, enable_filesystem_cache_on_write_operations = true, filesystem_cache_segments_batch_size = 5, use_page_cache_for_disks_without_file_cache = true, page_cache_inject_eviction = true, load_marks_asynchronously = true, allow_prefetched_read_pool_for_remote_filesystem = false, allow_prefetched_read_pool_for_local_filesystem = false, filesystem_prefetch_max_memory_usage = 33554432, filesystem_prefetches_limit = 10, max_streams_for_merge_tree_reading = 1000, optimize_sorting_by_input_stream_properties = false, insert_keeper_max_retries = 100, insert_keeper_retry_initial_backoff_ms = 1, insert_keeper_retry_max_backoff_ms = 10, insert_keeper_fault_injection_probability = 0.009999999776482582, optimize_distinct_in_order = false, allow_experimental_parallel_reading_from_replicas = 0, parallel_replicas_for_non_replicated_merge_tree = true, parallel_replicas_local_plan = false, session_timezone = 'Mexico/BajaSur', output_format_native_write_json_as_string = true
2025.04.01 17:45:53.785273 [ 423 ] {} Application: Child process was terminated by signal 6.
+ rg -A50 -Fa ============ /var/log/clickhouse-server/stderr.log
+ :
+ data_path_config=--path=/var/lib/clickhouse/
+ zstd --threads=0
+ [[ -n '' ]]
+ '[' 1 -ne 0 ']'
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.query_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.zookeeper_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ zstd --threads=0
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.trace_log format TSVWithNamesAndTypes'
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.transactions_info_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ zstd --threads=0
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.metric_log format TSVWithNamesAndTypes'
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ zstd --threads=0
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.blob_storage_log format TSVWithNamesAndTypes'
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.error_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.query_metric_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for trace_type in CPU Memory Real
+ clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q '
select
arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack,
count(*) AS samples
from system.trace_log
where trace_type = '\''CPU'\''
group by trace
order by samples desc
settings allow_introspection_functions = 1
format TabSeparated'
+ zstd --threads=0
+ for trace_type in CPU Memory Real
+ clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q '
select
arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack,
count(*) AS samples
from system.trace_log
where trace_type = '\''Memory'\''
group by trace
order by samples desc
settings allow_introspection_functions = 1
format TabSeparated'
+ zstd --threads=0
+ for trace_type in CPU Memory Real
+ clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q '
select
arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack,
count(*) AS samples
from system.trace_log
where trace_type = '\''Real'\''
group by trace
order by samples desc
settings allow_introspection_functions = 1
format TabSeparated'
+ zstd --threads=0
+ check_logs_for_critical_errors
+ sed -n '/WARNING:.*anitizer/,/^$/p' /var/log/clickhouse-server/stderr.log
+ rg -Fav -e 'ASan doesn'\''t fully support makecontext/swapcontext functions' -e DB::Exception /test_output/tmp
+ echo -e 'No sanitizer asserts\tOK\t\N\t'
+ rm -f /test_output/tmp
+ rg -Fa ' Application: Child process was terminated by signal 9' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ echo -e 'No OOM messages in clickhouse-server.log\tOK\t\N\t'
+ rg -Fa 'Code: 49. DB::Exception: ' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ echo -e 'No logical errors\tOK\t\N\t'
+ '[' -s /test_output/logical_errors.txt ']'
+ rm /test_output/logical_errors.txt
+ rg --text 'Code: 499.*The specified key does not exist' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ grep -v -e a.myext -e DistributedCacheTCPHandler -e ReadBufferFromDistributedCache -e ReadBufferFromS3 -e ReadBufferFromAzureBlobStorage -e AsynchronousBoundedReadBuffer -e 'caller id: None:DistribCache'
+ echo -e 'No lost s3 keys\tOK\t\N\t'
+ rg -Fa 'it is lost forever' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ grep SharedMergeTreePartCheckThread
+ echo -e 'No SharedMergeTree lost forever in clickhouse-server.log\tOK\t\N\t'
+ '[' -s /test_output/no_such_key_errors.txt ']'
+ rm /test_output/no_such_key_errors.txt
+ rg -Fa '#######################################' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ echo -e 'Killed by signal (in clickhouse-server.log)\tFAIL\t\N\t'
+ rg -Fa ' ' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
++ trim_server_logs fatal_messages.txt
++ grep -Eo ' \[ [0-9]+ \] \{.*'
++ head -n 100 /test_output/fatal_messages.txt
++ escaped
++ clickhouse local -S 's String' --input-format=LineAsString -q 'select substr(s, 1, 300)
from table format CustomSeparated settings format_custom_row_after_delimiter='\''\\\\n'\'''
+ echo -e 'Fatal message in clickhouse-server.log (see fatal_messages.txt)\tFAIL\t\N\t [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Logical error: \'\''Replica info is not initialized\'\''.\\n [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Stack trace (when copying this message, always include the lines below):\\n [ 40158 ] {} BaseDaemon: ########## Short fault info ############\\n [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2230) Received signal 6\\n [ 40158 ] {} BaseDaemon: Signal description: Aborted\\n [ 40158 ] {} BaseDaemon: \\n [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x0000\\n [ 40158 ] {} BaseDaemon: ########################################\\n [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2230) (query_id: 549c2d11-5172-44cf-9edd-d2571bb9f6a9) (query: INSERT INTO distributed_01099\\n [ 40158 ] {} BaseDaemon: \\n [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x0000\\n [ 40158 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()\\n [ 40158 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d\\n [ 40158 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422\\n [ 40158 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6\\n [ 40158 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb\\n [ 40158 ] {} BaseDaemon: 4. ? @ 0x00007f1bbdcda520\\n [ 40158 ] {} BaseDaemon: 5. ? @ 0x00007f1bbdd2e9fd\\n [ 40158 ] {} BaseDaemon: 6. ? @ 0x00007f1bbdcda476\\n [ 40158 ] {} BaseDaemon: 7. ? @ 0x00007f1bbdcc07f3\\n [ 40158 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267\\n [ 40158 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a\\n [ 40158 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)\\n [ 40158 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff\\n [ 40158 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade\\n [ 40158 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba\\n [ 40158 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c\\n [ 40158 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e\\n [ 40158 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d\\n [ 40158 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe\\n [ 40158 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b\\n [ 40158 ] {} BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d\\n [ 40158 ] {} BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*)\\n [ 40158 ] {} BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47\\n [ 40158 ] {} BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c\\n [ 40158 ] {} BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long)\\n [ 40158 ] {} BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc\\n [ 40158 ] {} BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa\\n [ 40158 ] {} BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool)\\n [ 40158 ] {} BaseDaemon: 23.1. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:89: operator()\\n [ 40158 ] {} BaseDaemon: 23.2. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?\\n [ 40158 ] {} BaseDaemon: 23.3. inlined from ./contrib/llvm-project/libcxx/include/tuple:1789: _ZNSt3__118__apply_tuple_implB6v15007IRZN2DB25CompletedPipelineExecutor7executeEvE3$_0RNS_5tupleIJEEETpTnmJEEEDcOT_OT0_NS_15__tuple_indicesIJXspT1_EEEE\\n [ 40158 ] {} BaseDaemon: 23.4. inlined from ./contrib/llvm-project/libcxx/include/tuple:1798: decltype(auto) std::apply[abi:v15007]&>(DB::CompletedPipelineExecutor::execute()::$_0&, std::tuple<>&)\\n [ 40158 ] {} BaseDaemon: 23.5. inlined from ./src/Common/ThreadPool.h:311: operator()\\n [ 40158 ] {} BaseDaemon: 23.6. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?\\n [ 40158 ] {} BaseDaemon: 23.7. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:479: ?\\n [ 40158 ] {} BaseDaemon: 23.8. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:235: ?\\n [ 40158 ] {} BaseDaemon: 23. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x000000001c52036a\\n [ 40158 ] {} BaseDaemon: 24.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:848: ?\\n [ 40158 ] {} BaseDaemon: 24.1. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:1197: ?\\n [ 40158 ] {} BaseDaemon: 24. ./build_docker/./src/Common/ThreadPool.cpp:785: ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x00000000104ef7ee\\n [ 40158 ] {} BaseDaemon: 25.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:359: ?\\n [ 40158 ] {} BaseDaemon: 25.1. inlined from ./contrib/llvm-project/libcxx/include/thread:284: void std::__thread_execute[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImp\\n [ 40158 ] {} BaseDaemon: 25. ./contrib/llvm-project/libcxx/include/thread:295: void* std::__thread_proxy[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl BaseDaemon: 26. __tsan_thread_start_func @ 0x0000000007b08428\\n [ 40158 ] {} BaseDaemon: 27. ? @ 0x00007f1bbdd2cac3\\n [ 40158 ] {} BaseDaemon: 28. ? @ 0x00007f1bbddbe850\\n [ 40158 ] {} BaseDaemon: Integrity check of the executable successfully passed (checksum: CAD5C347B3B532C05434FCAB6F84FB12)\\n [ 40158 ] {} BaseDaemon: This ClickHouse version is not official and should be upgraded to the official build.\\n [ 40158 ] {} BaseDaemon: Changed settings: min_compress_block_size = 838722, max_compress_block_size = 1814485, max_block_size = 43866, min_external_table_block_size_bytes = 0, max_joined_block_size_rows = 36602, max_insert_threads = 3, max_threads = 1, max_parsing_threads = 1, max_read_buf\\n [ 423 ] {} Application: Child process was terminated by signal 6.\\n [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Logical error: \'\''Replica info is not initialized\'\''.\\n [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Stack trace (when copying this message, always include the lines below):\\n [ 40158 ] {} BaseDaemon: ########## Short fault info ############\\n [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2230) Received signal 6\\n [ 40158 ] {} BaseDaemon: Signal description: Aborted\\n [ 40158 ] {} BaseDaemon: \\n [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x0000\\n [ 40158 ] {} BaseDaemon: ########################################\\n [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2230) (query_id: 549c2d11-5172-44cf-9edd-d2571bb9f6a9) (query: INSERT INTO distributed_01099\\n [ 40158 ] {} BaseDaemon: \\n [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x0000\\n [ 40158 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()\\n [ 40158 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d\\n [ 40158 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422\\n [ 40158 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6\\n [ 40158 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb\\n [ 40158 ] {} BaseDaemon: 4. ? @ 0x00007f1bbdcda520\\n [ 40158 ] {} BaseDaemon: 5. ? @ 0x00007f1bbdd2e9fd\\n [ 40158 ] {} BaseDaemon: 6. ? @ 0x00007f1bbdcda476\\n [ 40158 ] {} BaseDaemon: 7. ? @ 0x00007f1bbdcc07f3\\n [ 40158 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267\\n [ 40158 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a\\n [ 40158 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)\\n [ 40158 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff\\n [ 40158 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade\\n [ 40158 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba\\n [ 40158 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c\\n [ 40158 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e\\n [ 40158 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d\\n [ 40158 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe\\n [ 40158 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b\\n [ 40158 ] {} BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d\\n [ 40158 ] {} BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*)\\n [ 40158 ] {} BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47\\n [ 40158 ] {} BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c\\n [ 40158 ] {} BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long)\\n [ 40158 ] {} BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc\\n [ 40158 ] {} BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa\\n [ 40158 ] {} BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool)\\n'
+ '[' -s /test_output/fatal_messages.txt ']'
+ rg -v 'rg -Faz '
+ rg -Faz '########################################' /test_output/blob_storage_log.tsv.zst /test_output/clickhouse-server.log.zst /test_output/error_log.tsv.zst /test_output/fatal_messages.txt /test_output/gdb.log /test_output/metric_log.tsv.zst /test_output/minio.log /test_output/query_log.tsv.zst /test_output/query_metric_log.tsv.zst /test_output/run.log /test_output/test_results.tsv /test_output/test_result.txt /test_output/trace-log-CPU-flamegraph.tsv.zst /test_output/trace-log-Memory-flamegraph.tsv.zst /test_output/trace-log-Real-flamegraph.tsv.zst /test_output/trace_log.tsv.zst /test_output/transactions_info_log.tsv.zst /test_output/zookeeper_log.tsv.zst
+ echo -e 'Killed by signal (output files)\tFAIL\t\N\t'
+ rg -Fa ' received signal ' /test_output/gdb.log
+ dmesg -T
+ grep -q -F -e 'Out of memory: Killed process' -e 'oom_reaper: reaped process' -e oom-kill:constraint=CONSTRAINT_NONE /test_output/dmesg.log
+ echo -e 'No OOM in dmesg\tOK\t\N\t'
+ /repo/tests/docker_scripts/process_functional_tests_result.py
2025-04-01 17:59:55,658 File /repo/tests/broken_tests.json with broken tests found
2025-04-01 17:59:55,658 Find files in result folder trace-log-Memory-flamegraph.tsv.zst,fatal_messages.txt,trace-log-CPU-flamegraph.tsv.zst,clickhouse-server.log.zst,error_log.tsv.zst,test_result.txt,zookeeper_log.tsv.zst,query_metric_log.tsv.zst,gdb.log,trace_log.tsv.zst,test_results.tsv,dmesg.log,run.log,minio.log,query_log.tsv.zst,trace-log-Real-flamegraph.tsv.zst,blob_storage_log.tsv.zst,metric_log.tsv.zst,transactions_info_log.tsv.zst
2025-04-01 17:59:55,662 Is flaky check: False
2025-04-01 17:59:55,662 Result parsed
2025-04-01 17:59:55,664 Result written
File /repo/tests/analyzer_tech_debt.txt with broken tests found
Broken tests in the list: 3
Broken tests in the list: 3
+ rm /var/log/clickhouse-server/clickhouse-server.log
+ mv /var/log/clickhouse-server/stderr.log /test_output/
+ [[ -n '' ]]
+ tar -chf /test_output/coordination.tar /var/lib/clickhouse/coordination
tar: Removing leading `/' from member names
tar: Removing leading `/' from hard link targets
+ rm -rf /var/lib/clickhouse/data/system/asynchronous_insert_log/ /var/lib/clickhouse/data/system/asynchronous_metric_log/ /var/lib/clickhouse/data/system/backup_log/ /var/lib/clickhouse/data/system/blob_storage_log/ /var/lib/clickhouse/data/system/crash_log/ /var/lib/clickhouse/data/system/error_log/ /var/lib/clickhouse/data/system/filesystem_cache_log/ /var/lib/clickhouse/data/system/metric_log/ /var/lib/clickhouse/data/system/opentelemetry_span_log/ /var/lib/clickhouse/data/system/part_log/ /var/lib/clickhouse/data/system/processors_profile_log/ /var/lib/clickhouse/data/system/query_log/ /var/lib/clickhouse/data/system/query_metric_log/ /var/lib/clickhouse/data/system/query_thread_log/ /var/lib/clickhouse/data/system/query_views_log/ /var/lib/clickhouse/data/system/s3queue_log/ /var/lib/clickhouse/data/system/session_log/ /var/lib/clickhouse/data/system/text_log/ /var/lib/clickhouse/data/system/trace_log/ /var/lib/clickhouse/data/system/transactions_info_log/ /var/lib/clickhouse/data/system/zookeeper_log/
+ tar -chf /test_output/store.tar /var/lib/clickhouse/store
tar: Removing leading `/' from member names
tar: Removing leading `/' from hard link targets
+ tar -chf /test_output/metadata.tar /var/lib/clickhouse/metadata/default.sql /var/lib/clickhouse/metadata/information_schema.sql /var/lib/clickhouse/metadata/INFORMATION_SCHEMA.sql /var/lib/clickhouse/metadata/system.sql /var/lib/clickhouse/metadata/test_89ntz7bn.sql /var/lib/clickhouse/metadata/test_axigm1xv.sql /var/lib/clickhouse/metadata/test_jypz7w9c.sql /var/lib/clickhouse/metadata/test_l0nrtmbt.sql /var/lib/clickhouse/metadata/test_ofjicghs.sql /var/lib/clickhouse/metadata/test_ofr5rq7t.sql /var/lib/clickhouse/metadata/test_rnvzzqsk.sql /var/lib/clickhouse/metadata/test_sczjo540.sql /var/lib/clickhouse/metadata/test.sql /var/lib/clickhouse/metadata/test_umfejymz.sql
tar: Removing leading `/' from member names
tar: Removing leading `/' from hard link targets
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ collect_core_dumps
+ find . -type f -maxdepth 1 -name 'core.*'
+ read -r core